[ 518.108853] env[63202]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=63202) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 518.109242] env[63202]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=63202) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 518.109242] env[63202]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=63202) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 518.109578] env[63202]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 518.201233] env[63202]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=63202) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 518.211265] env[63202]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=63202) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 518.810213] env[63202]: INFO nova.virt.driver [None req-c7617c36-381b-466a-88bb-798052cff17c None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 518.881350] env[63202]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 518.881504] env[63202]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 518.881616] env[63202]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=63202) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 521.949746] env[63202]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-929f9410-8f69-4761-ae32-b8ce578d34e0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.966320] env[63202]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=63202) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 521.966513] env[63202]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-6ab21238-0b27-4571-b97a-b7f7c33208ae {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.999058] env[63202]: INFO oslo_vmware.api [-] Successfully established new session; session ID is 5990f. [ 521.999254] env[63202]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.118s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 521.999729] env[63202]: INFO nova.virt.vmwareapi.driver [None req-c7617c36-381b-466a-88bb-798052cff17c None None] VMware vCenter version: 7.0.3 [ 522.003151] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a06035cf-5e47-40bc-8d82-584d3de58483 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.020304] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c4b809e-626f-4054-bf1a-476ffaaded81 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.025780] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f154546-a447-43ae-98cf-c7827af3c039 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.032238] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b347f29-f41b-4435-b440-c61edb17984e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.044828] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88d6b654-1ed8-416b-a775-e6a69f890334 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.050401] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cbb7571-6d12-44f6-8bef-af1815aa053e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.080291] env[63202]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-1575bd05-96ee-4109-b38b-828815458809 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.085243] env[63202]: DEBUG nova.virt.vmwareapi.driver [None req-c7617c36-381b-466a-88bb-798052cff17c None None] Extension org.openstack.compute already exists. {{(pid=63202) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:226}} [ 522.087860] env[63202]: INFO nova.compute.provider_config [None req-c7617c36-381b-466a-88bb-798052cff17c None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 522.591528] env[63202]: DEBUG nova.context [None req-c7617c36-381b-466a-88bb-798052cff17c None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),97f9caa4-145d-43be-b385-f5d201c44331(cell1) {{(pid=63202) load_cells /opt/stack/nova/nova/context.py:464}} [ 522.593635] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 522.593945] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 522.594722] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 522.595193] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] Acquiring lock "97f9caa4-145d-43be-b385-f5d201c44331" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 522.595388] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] Lock "97f9caa4-145d-43be-b385-f5d201c44331" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 522.596594] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] Lock "97f9caa4-145d-43be-b385-f5d201c44331" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 522.617654] env[63202]: INFO dbcounter [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] Registered counter for database nova_cell0 [ 522.625926] env[63202]: INFO dbcounter [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] Registered counter for database nova_cell1 [ 522.629227] env[63202]: DEBUG oslo_db.sqlalchemy.engines [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=63202) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 522.629815] env[63202]: DEBUG oslo_db.sqlalchemy.engines [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=63202) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 522.634205] env[63202]: ERROR nova.db.main.api [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 522.634205] env[63202]: result = function(*args, **kwargs) [ 522.634205] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 522.634205] env[63202]: return func(*args, **kwargs) [ 522.634205] env[63202]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 522.634205] env[63202]: result = fn(*args, **kwargs) [ 522.634205] env[63202]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 522.634205] env[63202]: return f(*args, **kwargs) [ 522.634205] env[63202]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 522.634205] env[63202]: return db.service_get_minimum_version(context, binaries) [ 522.634205] env[63202]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 522.634205] env[63202]: _check_db_access() [ 522.634205] env[63202]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 522.634205] env[63202]: stacktrace = ''.join(traceback.format_stack()) [ 522.634205] env[63202]: [ 522.635316] env[63202]: ERROR nova.db.main.api [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 522.635316] env[63202]: result = function(*args, **kwargs) [ 522.635316] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 522.635316] env[63202]: return func(*args, **kwargs) [ 522.635316] env[63202]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 522.635316] env[63202]: result = fn(*args, **kwargs) [ 522.635316] env[63202]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 522.635316] env[63202]: return f(*args, **kwargs) [ 522.635316] env[63202]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 522.635316] env[63202]: return db.service_get_minimum_version(context, binaries) [ 522.635316] env[63202]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 522.635316] env[63202]: _check_db_access() [ 522.635316] env[63202]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 522.635316] env[63202]: stacktrace = ''.join(traceback.format_stack()) [ 522.635316] env[63202]: [ 522.635918] env[63202]: WARNING nova.objects.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 522.635918] env[63202]: WARNING nova.objects.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] Failed to get minimum service version for cell 97f9caa4-145d-43be-b385-f5d201c44331 [ 522.636288] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] Acquiring lock "singleton_lock" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 522.636451] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] Acquired lock "singleton_lock" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 522.636691] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] Releasing lock "singleton_lock" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 522.637020] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] Full set of CONF: {{(pid=63202) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 522.637164] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] ******************************************************************************** {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 522.637292] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] Configuration options gathered from: {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 522.637424] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 522.637612] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 522.637738] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] ================================================================================ {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 522.637979] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] allow_resize_to_same_host = True {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.638182] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] arq_binding_timeout = 300 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.638314] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] backdoor_port = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.638441] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] backdoor_socket = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.638605] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] block_device_allocate_retries = 60 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.638764] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] block_device_allocate_retries_interval = 3 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.638934] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cert = self.pem {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.639115] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.639287] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] compute_monitors = [] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.639455] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] config_dir = [] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.639627] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] config_drive_format = iso9660 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.639767] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.639937] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] config_source = [] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.640118] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] console_host = devstack {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.640285] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] control_exchange = nova {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.640444] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cpu_allocation_ratio = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.640606] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] daemon = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.640772] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] debug = True {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.640930] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] default_access_ip_network_name = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.641113] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] default_availability_zone = nova {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.641274] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] default_ephemeral_format = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.641436] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] default_green_pool_size = 1000 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.641671] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.641836] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] default_schedule_zone = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.641996] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] disk_allocation_ratio = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.642175] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] enable_new_services = True {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.642350] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] enabled_apis = ['osapi_compute'] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.642511] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] enabled_ssl_apis = [] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.642671] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] flat_injected = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.642827] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] force_config_drive = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.642984] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] force_raw_images = True {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.643168] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] graceful_shutdown_timeout = 5 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.643328] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] heal_instance_info_cache_interval = 60 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.643549] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] host = cpu-1 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.643710] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.643896] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] initial_disk_allocation_ratio = 1.0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.644081] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] initial_ram_allocation_ratio = 1.0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.644298] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.644467] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] instance_build_timeout = 0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.644632] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] instance_delete_interval = 300 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.644798] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] instance_format = [instance: %(uuid)s] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.644966] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] instance_name_template = instance-%08x {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.645147] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] instance_usage_audit = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.645317] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] instance_usage_audit_period = month {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.645480] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.645645] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] instances_path = /opt/stack/data/nova/instances {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.645809] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] internal_service_availability_zone = internal {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.645997] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] key = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.646181] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] live_migration_retry_count = 30 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.646353] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] log_color = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.646518] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] log_config_append = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.646682] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.646840] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] log_dir = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.646997] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] log_file = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.647140] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] log_options = True {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.647302] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] log_rotate_interval = 1 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.647469] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] log_rotate_interval_type = days {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.647635] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] log_rotation_type = none {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.647764] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.647891] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.648076] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.648247] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.648375] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.648538] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] long_rpc_timeout = 1800 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.648698] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] max_concurrent_builds = 10 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.648856] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] max_concurrent_live_migrations = 1 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.649022] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] max_concurrent_snapshots = 5 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.649187] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] max_local_block_devices = 3 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.649344] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] max_logfile_count = 30 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.649500] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] max_logfile_size_mb = 200 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.649656] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] maximum_instance_delete_attempts = 5 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.649819] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] metadata_listen = 0.0.0.0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.650037] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] metadata_listen_port = 8775 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.650227] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] metadata_workers = 2 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.650391] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] migrate_max_retries = -1 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.650555] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] mkisofs_cmd = genisoimage {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.650757] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] my_block_storage_ip = 10.180.1.21 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.650890] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] my_ip = 10.180.1.21 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.651066] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] network_allocate_retries = 0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.651251] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.651419] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] osapi_compute_listen = 0.0.0.0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.651581] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] osapi_compute_listen_port = 8774 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.651746] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] osapi_compute_unique_server_name_scope = {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.651913] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] osapi_compute_workers = 2 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.652118] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] password_length = 12 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.652302] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] periodic_enable = True {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.652465] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] periodic_fuzzy_delay = 60 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.652633] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] pointer_model = usbtablet {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.652798] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] preallocate_images = none {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.652959] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] publish_errors = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.653102] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] pybasedir = /opt/stack/nova {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.653264] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] ram_allocation_ratio = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.653424] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] rate_limit_burst = 0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.653643] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] rate_limit_except_level = CRITICAL {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.653748] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] rate_limit_interval = 0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.653975] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] reboot_timeout = 0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.654162] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] reclaim_instance_interval = 0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.654321] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] record = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.654490] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] reimage_timeout_per_gb = 60 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.654655] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] report_interval = 120 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.654814] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] rescue_timeout = 0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.654974] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] reserved_host_cpus = 0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.655149] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] reserved_host_disk_mb = 0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.655309] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] reserved_host_memory_mb = 512 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.655467] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] reserved_huge_pages = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.655627] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] resize_confirm_window = 0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.655786] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] resize_fs_using_block_device = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.655975] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] resume_guests_state_on_host_boot = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.656166] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.656330] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] rpc_response_timeout = 60 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.656490] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] run_external_periodic_tasks = True {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.656658] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] running_deleted_instance_action = reap {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.656816] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] running_deleted_instance_poll_interval = 1800 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.656994] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] running_deleted_instance_timeout = 0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.657173] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] scheduler_instance_sync_interval = 120 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.657342] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] service_down_time = 720 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.657509] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] servicegroup_driver = db {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.657663] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] shell_completion = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.657821] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] shelved_offload_time = 0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.657977] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] shelved_poll_interval = 3600 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.658161] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] shutdown_timeout = 0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.658324] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] source_is_ipv6 = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.658482] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] ssl_only = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.658727] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.658896] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] sync_power_state_interval = 600 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.659072] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] sync_power_state_pool_size = 1000 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.659244] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] syslog_log_facility = LOG_USER {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.659401] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] tempdir = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.659560] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] timeout_nbd = 10 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.659727] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] transport_url = **** {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.659889] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] update_resources_interval = 0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.660059] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] use_cow_images = True {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.660221] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] use_eventlog = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.660379] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] use_journal = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.660535] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] use_json = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.660691] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] use_rootwrap_daemon = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.660848] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] use_stderr = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.661055] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] use_syslog = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.661247] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vcpu_pin_set = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.661421] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vif_plugging_is_fatal = True {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.661585] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vif_plugging_timeout = 300 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.661749] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] virt_mkfs = [] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.661910] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] volume_usage_poll_interval = 0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.662132] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] watch_log_file = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.662315] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] web = /usr/share/spice-html5 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 522.662501] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] os_brick.lock_path = /opt/stack/data/n-cpu-1 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.662671] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.662837] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.663015] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_concurrency.disable_process_locking = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.663312] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.663492] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.663661] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.663831] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.664007] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.664184] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.664366] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] api.auth_strategy = keystone {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.664531] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] api.compute_link_prefix = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.664705] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.664875] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] api.dhcp_domain = novalocal {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.665058] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] api.enable_instance_password = True {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.665227] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] api.glance_link_prefix = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.665393] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.665566] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.665728] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] api.instance_list_per_project_cells = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.665904] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] api.list_records_by_skipping_down_cells = True {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.666101] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] api.local_metadata_per_cell = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.666278] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] api.max_limit = 1000 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.666445] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] api.metadata_cache_expiration = 15 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.666617] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] api.neutron_default_tenant_id = default {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.666787] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] api.response_validation = warn {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.666964] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] api.use_neutron_default_nets = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.667171] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.667341] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.667510] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.667684] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.667857] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] api.vendordata_dynamic_targets = [] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.668040] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] api.vendordata_jsonfile_path = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.668232] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.668421] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cache.backend = dogpile.cache.memcached {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.668589] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cache.backend_argument = **** {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.668756] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cache.config_prefix = cache.oslo {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.668924] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cache.dead_timeout = 60.0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.669119] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cache.debug_cache_backend = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.669298] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cache.enable_retry_client = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.669462] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cache.enable_socket_keepalive = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.669633] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cache.enabled = True {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.669797] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cache.enforce_fips_mode = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.669962] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cache.expiration_time = 600 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.670142] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cache.hashclient_retry_attempts = 2 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.670309] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cache.hashclient_retry_delay = 1.0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.670470] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cache.memcache_dead_retry = 300 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.670627] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cache.memcache_password = **** {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.670789] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.670950] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.671155] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cache.memcache_pool_maxsize = 10 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.671369] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.671542] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cache.memcache_sasl_enabled = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.671724] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.671893] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cache.memcache_socket_timeout = 1.0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.672079] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cache.memcache_username = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.672315] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cache.proxies = [] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.672496] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cache.redis_db = 0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.672659] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cache.redis_password = **** {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.672831] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cache.redis_sentinel_service_name = mymaster {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.673010] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.673192] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cache.redis_server = localhost:6379 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.673357] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cache.redis_socket_timeout = 1.0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.673519] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cache.redis_username = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.673680] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cache.retry_attempts = 2 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.673865] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cache.retry_delay = 0.0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.673992] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cache.socket_keepalive_count = 1 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.674167] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cache.socket_keepalive_idle = 1 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.674330] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cache.socket_keepalive_interval = 1 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.674486] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cache.tls_allowed_ciphers = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.674641] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cache.tls_cafile = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.674795] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cache.tls_certfile = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.674961] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cache.tls_enabled = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.675137] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cache.tls_keyfile = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.675305] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cinder.auth_section = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.675479] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cinder.auth_type = password {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.675640] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cinder.cafile = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.675813] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cinder.catalog_info = volumev3::publicURL {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.675996] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cinder.certfile = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.676184] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cinder.collect_timing = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.676346] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cinder.cross_az_attach = True {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.676508] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cinder.debug = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.676667] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cinder.endpoint_template = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.676830] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cinder.http_retries = 3 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.677015] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cinder.insecure = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.677197] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cinder.keyfile = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.677370] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cinder.os_region_name = RegionOne {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.677535] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cinder.split_loggers = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.677694] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cinder.timeout = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.677866] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.678039] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] compute.cpu_dedicated_set = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.678231] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] compute.cpu_shared_set = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.678403] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] compute.image_type_exclude_list = [] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.678568] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.678732] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] compute.max_concurrent_disk_ops = 0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.678894] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] compute.max_disk_devices_to_attach = -1 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.679069] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.679241] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.679402] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] compute.resource_provider_association_refresh = 300 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.679563] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.679724] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] compute.shutdown_retry_interval = 10 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.679905] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.680091] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] conductor.workers = 2 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.680273] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] console.allowed_origins = [] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.680434] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] console.ssl_ciphers = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.680601] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] console.ssl_minimum_version = default {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.680768] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] consoleauth.enforce_session_timeout = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.680936] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] consoleauth.token_ttl = 600 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.681141] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cyborg.cafile = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.681314] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cyborg.certfile = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.681479] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cyborg.collect_timing = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.681638] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cyborg.connect_retries = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.681799] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cyborg.connect_retry_delay = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.681955] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cyborg.endpoint_override = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.682135] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cyborg.insecure = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.682297] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cyborg.keyfile = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.682459] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cyborg.max_version = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.682618] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cyborg.min_version = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.682776] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cyborg.region_name = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.682935] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cyborg.retriable_status_codes = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.683104] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cyborg.service_name = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.683276] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cyborg.service_type = accelerator {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.683436] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cyborg.split_loggers = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.683596] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cyborg.status_code_retries = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.683755] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cyborg.status_code_retry_delay = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.683965] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cyborg.timeout = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.684156] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.684321] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] cyborg.version = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.684500] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] database.backend = sqlalchemy {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.684671] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] database.connection = **** {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.684836] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] database.connection_debug = 0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.685014] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] database.connection_parameters = {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.685190] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] database.connection_recycle_time = 3600 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.685353] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] database.connection_trace = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.685513] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] database.db_inc_retry_interval = True {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.685675] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] database.db_max_retries = 20 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.685837] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] database.db_max_retry_interval = 10 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.686037] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] database.db_retry_interval = 1 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.686212] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] database.max_overflow = 50 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.686378] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] database.max_pool_size = 5 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.686539] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] database.max_retries = 10 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.686708] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.686869] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] database.mysql_wsrep_sync_wait = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.687036] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] database.pool_timeout = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.687206] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] database.retry_interval = 10 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.687364] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] database.slave_connection = **** {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.687524] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] database.sqlite_synchronous = True {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.687684] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] database.use_db_reconnect = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.687860] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] api_database.backend = sqlalchemy {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.688037] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] api_database.connection = **** {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.688209] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] api_database.connection_debug = 0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.688379] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] api_database.connection_parameters = {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.688541] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] api_database.connection_recycle_time = 3600 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.688703] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] api_database.connection_trace = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.688864] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] api_database.db_inc_retry_interval = True {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.689037] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] api_database.db_max_retries = 20 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.689206] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] api_database.db_max_retry_interval = 10 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.689366] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] api_database.db_retry_interval = 1 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.689524] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] api_database.max_overflow = 50 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.689683] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] api_database.max_pool_size = 5 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.689842] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] api_database.max_retries = 10 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.690019] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.690183] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.690340] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] api_database.pool_timeout = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.690500] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] api_database.retry_interval = 10 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.690655] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] api_database.slave_connection = **** {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.690816] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] api_database.sqlite_synchronous = True {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.690990] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] devices.enabled_mdev_types = [] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.691184] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.691352] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] ephemeral_storage_encryption.default_format = luks {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.691515] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] ephemeral_storage_encryption.enabled = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.691676] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.691846] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] glance.api_servers = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.692020] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] glance.cafile = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.692181] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] glance.certfile = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.692345] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] glance.collect_timing = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.692503] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] glance.connect_retries = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.692661] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] glance.connect_retry_delay = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.692820] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] glance.debug = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.692984] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] glance.default_trusted_certificate_ids = [] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.693162] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] glance.enable_certificate_validation = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.693326] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] glance.enable_rbd_download = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.693484] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] glance.endpoint_override = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.693650] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] glance.insecure = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.693829] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] glance.keyfile = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.694062] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] glance.max_version = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.694185] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] glance.min_version = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.694352] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] glance.num_retries = 3 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.694522] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] glance.rbd_ceph_conf = {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.694686] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] glance.rbd_connect_timeout = 5 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.694852] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] glance.rbd_pool = {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.695031] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] glance.rbd_user = {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.695199] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] glance.region_name = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.695358] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] glance.retriable_status_codes = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.695514] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] glance.service_name = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.695679] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] glance.service_type = image {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.695858] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] glance.split_loggers = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.696079] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] glance.status_code_retries = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.696252] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] glance.status_code_retry_delay = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.696412] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] glance.timeout = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.696592] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.696757] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] glance.verify_glance_signatures = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.696918] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] glance.version = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.697101] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] guestfs.debug = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.697271] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] manila.auth_section = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.697433] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] manila.auth_type = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.697593] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] manila.cafile = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.697749] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] manila.certfile = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.697911] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] manila.collect_timing = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.698085] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] manila.connect_retries = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.698244] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] manila.connect_retry_delay = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.698400] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] manila.endpoint_override = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.698561] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] manila.insecure = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.698717] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] manila.keyfile = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.698874] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] manila.max_version = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.699045] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] manila.min_version = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.699207] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] manila.region_name = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.699366] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] manila.retriable_status_codes = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.699520] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] manila.service_name = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.699685] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] manila.service_type = shared-file-system {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.699845] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] manila.share_apply_policy_timeout = 10 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.700012] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] manila.split_loggers = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.700178] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] manila.status_code_retries = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.700335] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] manila.status_code_retry_delay = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.700491] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] manila.timeout = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.700669] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] manila.valid_interfaces = ['internal', 'public'] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.700828] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] manila.version = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.700998] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] mks.enabled = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.701363] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.701551] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] image_cache.manager_interval = 2400 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.701720] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] image_cache.precache_concurrency = 1 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.701889] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] image_cache.remove_unused_base_images = True {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.702072] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.702243] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.702416] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] image_cache.subdirectory_name = _base {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.702588] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] ironic.api_max_retries = 60 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.702752] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] ironic.api_retry_interval = 2 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.702910] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] ironic.auth_section = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.703088] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] ironic.auth_type = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.703255] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] ironic.cafile = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.703410] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] ironic.certfile = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.703571] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] ironic.collect_timing = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.703732] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] ironic.conductor_group = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.703935] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] ironic.connect_retries = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.704188] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] ironic.connect_retry_delay = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.704298] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] ironic.endpoint_override = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.704446] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] ironic.insecure = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.704600] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] ironic.keyfile = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.704758] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] ironic.max_version = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.704916] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] ironic.min_version = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.705094] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] ironic.peer_list = [] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.705258] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] ironic.region_name = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.705415] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] ironic.retriable_status_codes = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.705576] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] ironic.serial_console_state_timeout = 10 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.705731] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] ironic.service_name = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.705923] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] ironic.service_type = baremetal {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.706102] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] ironic.shard = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.706267] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] ironic.split_loggers = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.706426] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] ironic.status_code_retries = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.706584] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] ironic.status_code_retry_delay = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.706740] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] ironic.timeout = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.706920] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.707100] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] ironic.version = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.707285] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.707458] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] key_manager.fixed_key = **** {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.707640] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.707801] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] barbican.barbican_api_version = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.707998] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] barbican.barbican_endpoint = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.708198] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] barbican.barbican_endpoint_type = public {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.708361] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] barbican.barbican_region_name = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.708521] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] barbican.cafile = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.708679] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] barbican.certfile = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.708841] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] barbican.collect_timing = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.709010] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] barbican.insecure = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.709177] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] barbican.keyfile = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.709341] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] barbican.number_of_retries = 60 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.709503] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] barbican.retry_delay = 1 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.709665] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] barbican.send_service_user_token = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.709824] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] barbican.split_loggers = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.709981] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] barbican.timeout = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.710157] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] barbican.verify_ssl = True {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.710316] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] barbican.verify_ssl_path = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.710479] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] barbican_service_user.auth_section = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.710642] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] barbican_service_user.auth_type = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.710797] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] barbican_service_user.cafile = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.710952] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] barbican_service_user.certfile = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.711130] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] barbican_service_user.collect_timing = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.711291] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] barbican_service_user.insecure = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.711450] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] barbican_service_user.keyfile = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.711611] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] barbican_service_user.split_loggers = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.711768] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] barbican_service_user.timeout = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.711932] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vault.approle_role_id = **** {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.712103] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vault.approle_secret_id = **** {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.712276] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vault.kv_mountpoint = secret {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.712436] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vault.kv_path = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.712598] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vault.kv_version = 2 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.712755] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vault.namespace = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.712915] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vault.root_token_id = **** {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.713083] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vault.ssl_ca_crt_file = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.713251] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vault.timeout = 60.0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.713411] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vault.use_ssl = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.713576] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.713750] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] keystone.auth_section = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.713953] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] keystone.auth_type = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.714140] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] keystone.cafile = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.714305] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] keystone.certfile = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.714468] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] keystone.collect_timing = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.714627] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] keystone.connect_retries = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.714785] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] keystone.connect_retry_delay = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.714943] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] keystone.endpoint_override = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.715120] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] keystone.insecure = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.715278] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] keystone.keyfile = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.715434] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] keystone.max_version = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.715590] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] keystone.min_version = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.715744] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] keystone.region_name = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.715960] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] keystone.retriable_status_codes = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.716164] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] keystone.service_name = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.716355] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] keystone.service_type = identity {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.716582] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] keystone.split_loggers = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.716773] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] keystone.status_code_retries = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.717061] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] keystone.status_code_retry_delay = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.717280] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] keystone.timeout = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.717527] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.717701] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] keystone.version = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.717973] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.connection_uri = {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.718161] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.cpu_mode = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.718388] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.cpu_model_extra_flags = [] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.718575] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.cpu_models = [] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.718815] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.cpu_power_governor_high = performance {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.719091] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.cpu_power_governor_low = powersave {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.719335] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.cpu_power_management = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.719559] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.719764] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.device_detach_attempts = 8 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.719982] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.device_detach_timeout = 20 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.720207] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.disk_cachemodes = [] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.720396] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.disk_prefix = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.720619] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.enabled_perf_events = [] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.720802] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.file_backed_memory = 0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.721040] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.gid_maps = [] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.721216] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.hw_disk_discard = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.721431] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.hw_machine_type = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.721628] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.images_rbd_ceph_conf = {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.721809] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.722018] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.722230] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.images_rbd_glance_store_name = {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.722423] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.images_rbd_pool = rbd {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.722649] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.images_type = default {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.722829] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.images_volume_group = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.723047] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.inject_key = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.723236] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.inject_partition = -2 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.723456] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.inject_password = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.723644] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.iscsi_iface = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.723864] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.iser_use_multipath = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.724093] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.live_migration_bandwidth = 0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.724329] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.724514] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.live_migration_downtime = 500 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.724724] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.724924] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.725127] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.live_migration_inbound_addr = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.725340] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.725525] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.live_migration_permit_post_copy = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.725750] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.live_migration_scheme = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.725956] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.live_migration_timeout_action = abort {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.726196] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.live_migration_tunnelled = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.726382] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.live_migration_uri = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.726597] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.live_migration_with_native_tls = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.726793] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.max_queues = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.727035] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.727287] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.727496] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.nfs_mount_options = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.727796] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.727971] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.728159] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.num_iser_scan_tries = 5 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.728323] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.num_memory_encrypted_guests = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.728487] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.728650] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.num_pcie_ports = 0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.728818] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.num_volume_scan_tries = 5 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.728985] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.pmem_namespaces = [] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.729165] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.quobyte_client_cfg = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.729452] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.729625] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.rbd_connect_timeout = 5 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.729793] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.729958] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.730136] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.rbd_secret_uuid = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.730298] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.rbd_user = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.730462] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.730633] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.remote_filesystem_transport = ssh {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.730793] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.rescue_image_id = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.730951] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.rescue_kernel_id = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.731126] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.rescue_ramdisk_id = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.731295] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.731454] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.rx_queue_size = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.731618] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.smbfs_mount_options = {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.731889] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.732073] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.snapshot_compression = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.732239] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.snapshot_image_format = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.732458] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.732695] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.sparse_logical_volumes = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.732871] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.swtpm_enabled = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.733056] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.swtpm_group = tss {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.733228] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.swtpm_user = tss {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.733397] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.sysinfo_serial = unique {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.733556] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.tb_cache_size = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.733714] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.tx_queue_size = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.733876] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.uid_maps = [] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.734049] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.use_virtio_for_bridges = True {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.734223] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.virt_type = kvm {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.734387] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.volume_clear = zero {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.734553] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.volume_clear_size = 0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.734707] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.volume_use_multipath = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.734861] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.vzstorage_cache_path = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.735036] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.735206] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.vzstorage_mount_group = qemu {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.735367] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.vzstorage_mount_opts = [] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.735531] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.735800] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.735985] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.vzstorage_mount_user = stack {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.736162] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.736335] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] neutron.auth_section = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.736505] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] neutron.auth_type = password {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.736665] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] neutron.cafile = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.736820] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] neutron.certfile = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.736982] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] neutron.collect_timing = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.737156] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] neutron.connect_retries = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.737314] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] neutron.connect_retry_delay = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.737478] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] neutron.default_floating_pool = public {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.737634] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] neutron.endpoint_override = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.737791] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] neutron.extension_sync_interval = 600 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.737953] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] neutron.http_retries = 3 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.738129] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] neutron.insecure = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.738290] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] neutron.keyfile = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.738450] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] neutron.max_version = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.738619] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.738775] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] neutron.min_version = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.738940] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] neutron.ovs_bridge = br-int {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.739119] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] neutron.physnets = [] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.739290] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] neutron.region_name = RegionOne {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.739449] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] neutron.retriable_status_codes = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.739615] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] neutron.service_metadata_proxy = True {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.739771] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] neutron.service_name = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.739934] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] neutron.service_type = network {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.740109] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] neutron.split_loggers = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.740270] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] neutron.status_code_retries = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.740428] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] neutron.status_code_retry_delay = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.740585] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] neutron.timeout = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.740762] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.740922] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] neutron.version = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.741108] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] notifications.bdms_in_notifications = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.741292] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] notifications.default_level = INFO {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.741465] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] notifications.notification_format = unversioned {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.741627] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] notifications.notify_on_state_change = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.741801] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.741974] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] pci.alias = [] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.742155] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] pci.device_spec = [] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.742321] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] pci.report_in_placement = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.742490] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] placement.auth_section = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.742658] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] placement.auth_type = password {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.742823] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.742982] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] placement.cafile = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.743154] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] placement.certfile = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.743317] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] placement.collect_timing = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.743474] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] placement.connect_retries = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.743630] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] placement.connect_retry_delay = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.743787] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] placement.default_domain_id = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.743943] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] placement.default_domain_name = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.744116] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] placement.domain_id = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.744273] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] placement.domain_name = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.744429] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] placement.endpoint_override = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.744586] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] placement.insecure = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.744743] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] placement.keyfile = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.744895] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] placement.max_version = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.745061] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] placement.min_version = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.745232] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] placement.password = **** {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.745387] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] placement.project_domain_id = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.745550] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] placement.project_domain_name = Default {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.745713] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] placement.project_id = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.745880] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] placement.project_name = service {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.746054] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] placement.region_name = RegionOne {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.746217] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] placement.retriable_status_codes = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.746373] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] placement.service_name = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.746538] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] placement.service_type = placement {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.746697] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] placement.split_loggers = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.746855] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] placement.status_code_retries = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.747016] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] placement.status_code_retry_delay = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.747176] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] placement.system_scope = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.747334] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] placement.timeout = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.747488] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] placement.trust_id = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.747643] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] placement.user_domain_id = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.747810] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] placement.user_domain_name = Default {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.747972] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] placement.user_id = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.748160] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] placement.username = nova {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.748341] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.748499] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] placement.version = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.748675] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] quota.cores = 20 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.748838] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] quota.count_usage_from_placement = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.749013] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.749196] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] quota.injected_file_content_bytes = 10240 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.749363] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] quota.injected_file_path_length = 255 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.749526] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] quota.injected_files = 5 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.749691] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] quota.instances = 10 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.749857] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] quota.key_pairs = 100 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.750033] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] quota.metadata_items = 128 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.750206] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] quota.ram = 51200 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.750370] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] quota.recheck_quota = True {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.750533] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] quota.server_group_members = 10 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.750695] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] quota.server_groups = 10 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.750864] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.751038] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.751206] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] scheduler.image_metadata_prefilter = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.751366] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.751526] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] scheduler.max_attempts = 3 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.751684] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] scheduler.max_placement_results = 1000 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.751841] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.752011] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] scheduler.query_placement_for_image_type_support = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.752175] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.752345] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] scheduler.workers = 2 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.752512] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.752680] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.752854] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.753030] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.753202] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.753367] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.753523] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.753705] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.753871] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] filter_scheduler.host_subset_size = 1 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.754046] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.754208] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.754368] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.754528] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] filter_scheduler.isolated_hosts = [] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.754691] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] filter_scheduler.isolated_images = [] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.754845] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.755008] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.755176] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.755334] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] filter_scheduler.pci_in_placement = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.755491] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.755647] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.755804] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.755982] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.756160] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.756324] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.756480] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] filter_scheduler.track_instance_changes = True {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.756651] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.756815] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] metrics.required = True {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.756976] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] metrics.weight_multiplier = 1.0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.757154] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.757315] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] metrics.weight_setting = [] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.757623] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.757803] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] serial_console.enabled = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.757980] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] serial_console.port_range = 10000:20000 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.758166] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.758336] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.758498] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] serial_console.serialproxy_port = 6083 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.758663] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] service_user.auth_section = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.758831] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] service_user.auth_type = password {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.758992] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] service_user.cafile = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.759167] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] service_user.certfile = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.759330] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] service_user.collect_timing = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.759490] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] service_user.insecure = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.759646] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] service_user.keyfile = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.759815] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] service_user.send_service_user_token = True {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.759976] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] service_user.split_loggers = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.760150] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] service_user.timeout = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.760319] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] spice.agent_enabled = True {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.760479] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] spice.enabled = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.760778] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.760967] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.761154] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] spice.html5proxy_port = 6082 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.761314] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] spice.image_compression = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.761471] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] spice.jpeg_compression = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.761626] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] spice.playback_compression = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.761783] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] spice.require_secure = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.761947] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] spice.server_listen = 127.0.0.1 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.762129] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.762288] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] spice.streaming_mode = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.762442] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] spice.zlib_compression = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.762604] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] upgrade_levels.baseapi = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.762768] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] upgrade_levels.compute = auto {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.762926] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] upgrade_levels.conductor = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.763093] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] upgrade_levels.scheduler = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.763261] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vendordata_dynamic_auth.auth_section = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.763423] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vendordata_dynamic_auth.auth_type = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.763578] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vendordata_dynamic_auth.cafile = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.763733] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vendordata_dynamic_auth.certfile = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.763892] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.764065] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vendordata_dynamic_auth.insecure = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.764223] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vendordata_dynamic_auth.keyfile = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.764382] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.764537] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vendordata_dynamic_auth.timeout = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.764706] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vmware.api_retry_count = 10 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.764882] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vmware.ca_file = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.765040] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vmware.cache_prefix = devstack-image-cache {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.765212] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vmware.cluster_name = testcl1 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.765373] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vmware.connection_pool_size = 10 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.765532] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vmware.console_delay_seconds = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.765699] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vmware.datastore_regex = ^datastore.* {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.765906] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.766089] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vmware.host_password = **** {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.766260] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vmware.host_port = 443 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.766427] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vmware.host_username = administrator@vsphere.local {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.766596] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vmware.insecure = True {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.766757] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vmware.integration_bridge = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.766921] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vmware.maximum_objects = 100 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.767092] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vmware.pbm_default_policy = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.767257] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vmware.pbm_enabled = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.767413] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vmware.pbm_wsdl_location = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.767576] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.767732] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vmware.serial_port_proxy_uri = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.767886] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vmware.serial_port_service_uri = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.768060] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vmware.task_poll_interval = 0.5 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.768237] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vmware.use_linked_clone = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.768402] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vmware.vnc_keymap = en-us {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.768565] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vmware.vnc_port = 5900 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.768726] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vmware.vnc_port_total = 10000 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.768910] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vnc.auth_schemes = ['none'] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.769097] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vnc.enabled = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.769393] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.769575] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.769744] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vnc.novncproxy_port = 6080 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.769920] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vnc.server_listen = 127.0.0.1 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.770107] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.770271] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vnc.vencrypt_ca_certs = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.770426] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vnc.vencrypt_client_cert = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.770583] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vnc.vencrypt_client_key = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.770752] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.770912] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] workarounds.disable_deep_image_inspection = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.771086] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.771249] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.771407] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.771568] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] workarounds.disable_rootwrap = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.771725] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] workarounds.enable_numa_live_migration = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.771883] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.772053] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.772215] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.772374] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] workarounds.libvirt_disable_apic = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.772529] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.772687] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.772844] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.773019] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.773181] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.773338] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.773498] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.773656] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.773808] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.773969] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.774166] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.774334] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] wsgi.client_socket_timeout = 900 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.774498] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] wsgi.default_pool_size = 1000 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.774664] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] wsgi.keep_alive = True {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.774830] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] wsgi.max_header_line = 16384 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.774983] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] wsgi.secure_proxy_ssl_header = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.775161] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] wsgi.ssl_ca_file = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.775323] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] wsgi.ssl_cert_file = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.775480] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] wsgi.ssl_key_file = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.775643] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] wsgi.tcp_keepidle = 600 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.775817] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.775987] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] zvm.ca_file = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.776161] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] zvm.cloud_connector_url = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.776978] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.776978] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] zvm.reachable_timeout = 300 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.777189] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_policy.enforce_new_defaults = True {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.777564] env[63202]: WARNING oslo_config.cfg [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 522.777751] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_policy.enforce_scope = True {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.777931] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_policy.policy_default_rule = default {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.778132] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.778308] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_policy.policy_file = policy.yaml {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.778480] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.778639] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.778799] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.778957] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.779133] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.779305] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.779481] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.779657] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] profiler.connection_string = messaging:// {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.779823] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] profiler.enabled = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.779991] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] profiler.es_doc_type = notification {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.780174] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] profiler.es_scroll_size = 10000 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.780341] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] profiler.es_scroll_time = 2m {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.780500] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] profiler.filter_error_trace = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.780665] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] profiler.hmac_keys = **** {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.780829] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] profiler.sentinel_service_name = mymaster {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.780992] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] profiler.socket_timeout = 0.1 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.781172] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] profiler.trace_requests = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.781332] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] profiler.trace_sqlalchemy = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.781510] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] profiler_jaeger.process_tags = {} {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.781671] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] profiler_jaeger.service_name_prefix = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.781830] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] profiler_otlp.service_name_prefix = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.781991] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] remote_debug.host = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.782166] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] remote_debug.port = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.782342] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.782504] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.782664] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.782830] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.782983] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.783162] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.783325] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.783489] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.783647] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.783815] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.783973] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.784161] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.784329] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.784497] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.784664] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.784829] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.784995] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.785188] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.785352] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.785516] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.785679] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.785844] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.786042] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.786223] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.786388] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.786550] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.786713] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.786882] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.787082] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.787253] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_rabbit.ssl = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.787425] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.787595] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.787758] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.787927] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.788112] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_rabbit.ssl_version = {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.788277] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.788464] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.788631] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_notifications.retry = -1 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.788812] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.788988] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_messaging_notifications.transport_url = **** {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.789177] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_limit.auth_section = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.789342] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_limit.auth_type = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.789499] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_limit.cafile = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.789656] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_limit.certfile = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.789819] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_limit.collect_timing = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.789978] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_limit.connect_retries = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.790155] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_limit.connect_retry_delay = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.790314] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_limit.endpoint_id = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.790471] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_limit.endpoint_override = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.790632] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_limit.insecure = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.790790] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_limit.keyfile = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.790946] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_limit.max_version = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.791120] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_limit.min_version = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.791278] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_limit.region_name = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.791436] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_limit.retriable_status_codes = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.791594] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_limit.service_name = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.791748] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_limit.service_type = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.791907] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_limit.split_loggers = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.792075] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_limit.status_code_retries = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.792237] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_limit.status_code_retry_delay = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.792393] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_limit.timeout = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.792548] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_limit.valid_interfaces = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.792701] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_limit.version = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.792863] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_reports.file_event_handler = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.793038] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.793203] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] oslo_reports.log_dir = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.793373] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.793531] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.793688] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.793849] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.794016] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.794178] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.794349] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.794509] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vif_plug_ovs_privileged.group = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.794662] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.794823] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.794983] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.795181] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] vif_plug_ovs_privileged.user = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.795322] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] os_vif_linux_bridge.flat_interface = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.795498] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.795667] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.795833] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.796033] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.796208] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.796376] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.796537] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.796719] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.796905] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] os_vif_ovs.isolate_vif = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.797103] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.797275] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.797443] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.797609] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] os_vif_ovs.ovsdb_interface = native {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.797769] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] os_vif_ovs.per_port_bridge = False {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.797939] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] privsep_osbrick.capabilities = [21] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.798113] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] privsep_osbrick.group = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.798273] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] privsep_osbrick.helper_command = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.798436] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.798596] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.798752] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] privsep_osbrick.user = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.798922] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.799094] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] nova_sys_admin.group = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.799253] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] nova_sys_admin.helper_command = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.799413] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.799572] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.799724] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] nova_sys_admin.user = None {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 522.799848] env[63202]: DEBUG oslo_service.service [None req-b0b32220-7129-44b0-9b40-f5bfe746e5f3 None None] ******************************************************************************** {{(pid=63202) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 522.800327] env[63202]: INFO nova.service [-] Starting compute node (version 0.0.1) [ 523.303321] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a580d276-9ffc-49e7-b42b-821d5105f9d5 None None] Getting list of instances from cluster (obj){ [ 523.303321] env[63202]: value = "domain-c8" [ 523.303321] env[63202]: _type = "ClusterComputeResource" [ 523.303321] env[63202]: } {{(pid=63202) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 523.304518] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12aacc87-4293-43c3-bffd-7905f63c12eb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.313547] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a580d276-9ffc-49e7-b42b-821d5105f9d5 None None] Got total of 0 instances {{(pid=63202) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 523.314112] env[63202]: WARNING nova.virt.vmwareapi.driver [None req-a580d276-9ffc-49e7-b42b-821d5105f9d5 None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 523.314585] env[63202]: INFO nova.virt.node [None req-a580d276-9ffc-49e7-b42b-821d5105f9d5 None None] Generated node identity 79b33d17-3e75-494c-a550-67b275de2079 [ 523.314818] env[63202]: INFO nova.virt.node [None req-a580d276-9ffc-49e7-b42b-821d5105f9d5 None None] Wrote node identity 79b33d17-3e75-494c-a550-67b275de2079 to /opt/stack/data/n-cpu-1/compute_id [ 523.817977] env[63202]: WARNING nova.compute.manager [None req-a580d276-9ffc-49e7-b42b-821d5105f9d5 None None] Compute nodes ['79b33d17-3e75-494c-a550-67b275de2079'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 524.824137] env[63202]: INFO nova.compute.manager [None req-a580d276-9ffc-49e7-b42b-821d5105f9d5 None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 525.830341] env[63202]: WARNING nova.compute.manager [None req-a580d276-9ffc-49e7-b42b-821d5105f9d5 None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 525.830708] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a580d276-9ffc-49e7-b42b-821d5105f9d5 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.830822] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a580d276-9ffc-49e7-b42b-821d5105f9d5 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 525.830970] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a580d276-9ffc-49e7-b42b-821d5105f9d5 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 525.831145] env[63202]: DEBUG nova.compute.resource_tracker [None req-a580d276-9ffc-49e7-b42b-821d5105f9d5 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63202) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 525.832082] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad97dd3a-f0d8-434b-9e87-48841d2602fb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.840340] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a1ac0ed-0e64-4495-be63-bdb3bae33ad6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.855608] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd16785d-a7a7-437e-8436-9b75eed98e32 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.862206] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72697b47-3f60-4006-8e8a-ddc3723d95cf {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.890854] env[63202]: DEBUG nova.compute.resource_tracker [None req-a580d276-9ffc-49e7-b42b-821d5105f9d5 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181298MB free_disk=176GB free_vcpus=48 pci_devices=None {{(pid=63202) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 525.891022] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a580d276-9ffc-49e7-b42b-821d5105f9d5 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.891200] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a580d276-9ffc-49e7-b42b-821d5105f9d5 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 526.393657] env[63202]: WARNING nova.compute.resource_tracker [None req-a580d276-9ffc-49e7-b42b-821d5105f9d5 None None] No compute node record for cpu-1:79b33d17-3e75-494c-a550-67b275de2079: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 79b33d17-3e75-494c-a550-67b275de2079 could not be found. [ 526.900512] env[63202]: INFO nova.compute.resource_tracker [None req-a580d276-9ffc-49e7-b42b-821d5105f9d5 None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 79b33d17-3e75-494c-a550-67b275de2079 [ 528.409308] env[63202]: DEBUG nova.compute.resource_tracker [None req-a580d276-9ffc-49e7-b42b-821d5105f9d5 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=63202) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 528.409716] env[63202]: DEBUG nova.compute.resource_tracker [None req-a580d276-9ffc-49e7-b42b-821d5105f9d5 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=63202) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 528.564162] env[63202]: INFO nova.scheduler.client.report [None req-a580d276-9ffc-49e7-b42b-821d5105f9d5 None None] [req-62a99c2a-0928-4088-bd2f-ab36aeb07689] Created resource provider record via placement API for resource provider with UUID 79b33d17-3e75-494c-a550-67b275de2079 and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 528.577745] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82841e78-67ea-4b15-8b64-e08344ffb8d4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.585339] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-040b3397-c6fb-419f-88eb-83f1f82a49d3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.614645] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a8c9b33-39ca-4f41-af11-4aefb5d51d9b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.621430] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc7121ac-8107-45ed-9580-4da9d8cf7aee {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.633848] env[63202]: DEBUG nova.compute.provider_tree [None req-a580d276-9ffc-49e7-b42b-821d5105f9d5 None None] Updating inventory in ProviderTree for provider 79b33d17-3e75-494c-a550-67b275de2079 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 529.173573] env[63202]: DEBUG nova.scheduler.client.report [None req-a580d276-9ffc-49e7-b42b-821d5105f9d5 None None] Updated inventory for provider 79b33d17-3e75-494c-a550-67b275de2079 with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 529.173786] env[63202]: DEBUG nova.compute.provider_tree [None req-a580d276-9ffc-49e7-b42b-821d5105f9d5 None None] Updating resource provider 79b33d17-3e75-494c-a550-67b275de2079 generation from 0 to 1 during operation: update_inventory {{(pid=63202) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 529.173928] env[63202]: DEBUG nova.compute.provider_tree [None req-a580d276-9ffc-49e7-b42b-821d5105f9d5 None None] Updating inventory in ProviderTree for provider 79b33d17-3e75-494c-a550-67b275de2079 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 529.226929] env[63202]: DEBUG nova.compute.provider_tree [None req-a580d276-9ffc-49e7-b42b-821d5105f9d5 None None] Updating resource provider 79b33d17-3e75-494c-a550-67b275de2079 generation from 1 to 2 during operation: update_traits {{(pid=63202) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 529.730781] env[63202]: DEBUG nova.compute.resource_tracker [None req-a580d276-9ffc-49e7-b42b-821d5105f9d5 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63202) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 529.730781] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a580d276-9ffc-49e7-b42b-821d5105f9d5 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.837s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 529.730781] env[63202]: DEBUG nova.service [None req-a580d276-9ffc-49e7-b42b-821d5105f9d5 None None] Creating RPC server for service compute {{(pid=63202) start /opt/stack/nova/nova/service.py:186}} [ 529.743321] env[63202]: DEBUG nova.service [None req-a580d276-9ffc-49e7-b42b-821d5105f9d5 None None] Join ServiceGroup membership for this service compute {{(pid=63202) start /opt/stack/nova/nova/service.py:203}} [ 529.744061] env[63202]: DEBUG nova.servicegroup.drivers.db [None req-a580d276-9ffc-49e7-b42b-821d5105f9d5 None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=63202) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 546.749343] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._sync_power_states {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 547.252680] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Getting list of instances from cluster (obj){ [ 547.252680] env[63202]: value = "domain-c8" [ 547.252680] env[63202]: _type = "ClusterComputeResource" [ 547.252680] env[63202]: } {{(pid=63202) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 547.253893] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dad6e0c-9b80-477c-a6f2-a1f8104e7975 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.262720] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Got total of 0 instances {{(pid=63202) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 547.262982] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 547.263282] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Getting list of instances from cluster (obj){ [ 547.263282] env[63202]: value = "domain-c8" [ 547.263282] env[63202]: _type = "ClusterComputeResource" [ 547.263282] env[63202]: } {{(pid=63202) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 547.264108] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c963018d-d01e-4543-a56d-da8551aecbd7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.270944] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Got total of 0 instances {{(pid=63202) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 559.965377] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Acquiring lock "65cd7042-daaf-4845-a0b8-5d2d0be02696" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 559.965377] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Lock "65cd7042-daaf-4845-a0b8-5d2d0be02696" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.468043] env[63202]: DEBUG nova.compute.manager [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 561.014929] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.014929] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.002s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.016801] env[63202]: INFO nova.compute.claims [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 562.080634] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b60ca029-83ed-4e00-8030-40e47c9ec6c4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.093068] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daa23214-3fd2-409b-b2aa-8ffacecd795b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.139897] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02d8ec3a-00d3-4ae1-8463-15a16dcaa9d0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.149508] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-870299fa-5fdf-4779-907f-881865efeaf7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.165861] env[63202]: DEBUG nova.compute.provider_tree [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 562.670300] env[63202]: DEBUG nova.scheduler.client.report [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 563.176546] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.162s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 563.177675] env[63202]: DEBUG nova.compute.manager [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 563.685828] env[63202]: DEBUG nova.compute.utils [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 563.687403] env[63202]: DEBUG nova.compute.manager [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 563.687680] env[63202]: DEBUG nova.network.neutron [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 564.195727] env[63202]: DEBUG nova.compute.manager [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 565.206593] env[63202]: DEBUG nova.compute.manager [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 565.321623] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Acquiring lock "0ded82aa-c68d-4d11-9e4d-90ed9a19b708" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.321875] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Lock "0ded82aa-c68d-4d11-9e4d-90ed9a19b708" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.004s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.826173] env[63202]: DEBUG nova.compute.manager [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 565.994961] env[63202]: DEBUG nova.virt.hardware [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 565.995230] env[63202]: DEBUG nova.virt.hardware [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 565.995383] env[63202]: DEBUG nova.virt.hardware [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 565.995692] env[63202]: DEBUG nova.virt.hardware [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 565.995867] env[63202]: DEBUG nova.virt.hardware [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 565.996071] env[63202]: DEBUG nova.virt.hardware [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 565.996362] env[63202]: DEBUG nova.virt.hardware [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 565.996571] env[63202]: DEBUG nova.virt.hardware [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 565.998222] env[63202]: DEBUG nova.virt.hardware [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 565.998222] env[63202]: DEBUG nova.virt.hardware [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 565.998222] env[63202]: DEBUG nova.virt.hardware [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 565.998818] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72d07b41-b6c1-4914-bd6f-634906b5380b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.015514] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02f42545-c2e0-48e1-954d-b7a7819454f1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.040630] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28f6df78-6d48-46de-95ce-14e69485ff89 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.122621] env[63202]: DEBUG nova.policy [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '84347a46aa0e46c688f0dfc8b88f2534', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a734788bc5604560b29865b950555f7e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 566.358248] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.358672] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.361892] env[63202]: INFO nova.compute.claims [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 566.589293] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] Acquiring lock "d87b8f05-e3d6-4800-a635-af5b6f03373d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.589293] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] Lock "d87b8f05-e3d6-4800-a635-af5b6f03373d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.812176] env[63202]: DEBUG nova.network.neutron [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] Successfully created port: 4cf12941-cfc8-4fbf-9cd0-12b2427b08ba {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 567.096845] env[63202]: DEBUG nova.compute.manager [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 567.438720] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8daa7555-3520-483b-8e0a-080157da96ca {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.446950] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26c50aaa-2197-476b-aeb9-bc663ec9ba8b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.482544] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bda5d014-8cda-4113-afa0-650ff8ebdcbf {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.491954] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-120bda26-f956-400b-acf5-6539c63d4783 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.508637] env[63202]: DEBUG nova.compute.provider_tree [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 567.619385] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 568.016020] env[63202]: DEBUG nova.scheduler.client.report [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 568.526390] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.164s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 568.526390] env[63202]: DEBUG nova.compute.manager [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 568.529899] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.911s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 568.534131] env[63202]: INFO nova.compute.claims [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 568.912240] env[63202]: ERROR nova.compute.manager [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4cf12941-cfc8-4fbf-9cd0-12b2427b08ba, please check neutron logs for more information. [ 568.912240] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 568.912240] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 568.912240] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 568.912240] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 568.912240] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 568.912240] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 568.912240] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 568.912240] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 568.912240] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 568.912240] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 568.912240] env[63202]: ERROR nova.compute.manager raise self.value [ 568.912240] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 568.912240] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 568.912240] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 568.912240] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 568.913094] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 568.913094] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 568.913094] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4cf12941-cfc8-4fbf-9cd0-12b2427b08ba, please check neutron logs for more information. [ 568.913094] env[63202]: ERROR nova.compute.manager [ 568.913094] env[63202]: Traceback (most recent call last): [ 568.913094] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 568.913094] env[63202]: listener.cb(fileno) [ 568.913094] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 568.913094] env[63202]: result = function(*args, **kwargs) [ 568.913094] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 568.913094] env[63202]: return func(*args, **kwargs) [ 568.913094] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 568.913094] env[63202]: raise e [ 568.913094] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 568.913094] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 568.913094] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 568.913094] env[63202]: created_port_ids = self._update_ports_for_instance( [ 568.913094] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 568.913094] env[63202]: with excutils.save_and_reraise_exception(): [ 568.913094] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 568.913094] env[63202]: self.force_reraise() [ 568.913094] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 568.913094] env[63202]: raise self.value [ 568.913094] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 568.913094] env[63202]: updated_port = self._update_port( [ 568.913094] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 568.913094] env[63202]: _ensure_no_port_binding_failure(port) [ 568.913094] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 568.913094] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 568.913794] env[63202]: nova.exception.PortBindingFailed: Binding failed for port 4cf12941-cfc8-4fbf-9cd0-12b2427b08ba, please check neutron logs for more information. [ 568.913794] env[63202]: Removing descriptor: 14 [ 568.915413] env[63202]: ERROR nova.compute.manager [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4cf12941-cfc8-4fbf-9cd0-12b2427b08ba, please check neutron logs for more information. [ 568.915413] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] Traceback (most recent call last): [ 568.915413] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 568.915413] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] yield resources [ 568.915413] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 568.915413] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] self.driver.spawn(context, instance, image_meta, [ 568.915413] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 568.915413] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] self._vmops.spawn(context, instance, image_meta, injected_files, [ 568.915413] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 568.915413] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] vm_ref = self.build_virtual_machine(instance, [ 568.915413] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 568.915726] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] vif_infos = vmwarevif.get_vif_info(self._session, [ 568.915726] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 568.915726] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] for vif in network_info: [ 568.915726] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 568.915726] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] return self._sync_wrapper(fn, *args, **kwargs) [ 568.915726] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 568.915726] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] self.wait() [ 568.915726] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 568.915726] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] self[:] = self._gt.wait() [ 568.915726] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 568.915726] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] return self._exit_event.wait() [ 568.915726] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 568.915726] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] result = hub.switch() [ 568.916050] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 568.916050] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] return self.greenlet.switch() [ 568.916050] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 568.916050] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] result = function(*args, **kwargs) [ 568.916050] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 568.916050] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] return func(*args, **kwargs) [ 568.916050] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 568.916050] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] raise e [ 568.916050] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 568.916050] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] nwinfo = self.network_api.allocate_for_instance( [ 568.916050] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 568.916050] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] created_port_ids = self._update_ports_for_instance( [ 568.916050] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 568.916347] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] with excutils.save_and_reraise_exception(): [ 568.916347] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 568.916347] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] self.force_reraise() [ 568.916347] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 568.916347] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] raise self.value [ 568.916347] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 568.916347] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] updated_port = self._update_port( [ 568.916347] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 568.916347] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] _ensure_no_port_binding_failure(port) [ 568.916347] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 568.916347] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] raise exception.PortBindingFailed(port_id=port['id']) [ 568.916347] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] nova.exception.PortBindingFailed: Binding failed for port 4cf12941-cfc8-4fbf-9cd0-12b2427b08ba, please check neutron logs for more information. [ 568.916347] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] [ 568.916660] env[63202]: INFO nova.compute.manager [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] Terminating instance [ 568.918631] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Acquiring lock "refresh_cache-65cd7042-daaf-4845-a0b8-5d2d0be02696" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 568.918864] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Acquired lock "refresh_cache-65cd7042-daaf-4845-a0b8-5d2d0be02696" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 568.918933] env[63202]: DEBUG nova.network.neutron [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 569.039811] env[63202]: DEBUG nova.compute.utils [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 569.041202] env[63202]: DEBUG nova.compute.manager [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Not allocating networking since 'none' was specified. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 569.547650] env[63202]: DEBUG nova.compute.manager [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 569.576663] env[63202]: DEBUG nova.network.neutron [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 569.638214] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5dea846-3f17-46c7-aa6b-c74109835583 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.646583] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16fcc508-45e4-457d-8b2e-e7da75cd0551 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.682667] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49e1799a-376f-4378-83f2-c8621b22dd50 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.691566] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f89fff3b-92ec-42cc-842a-1b2af40a734a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.709871] env[63202]: DEBUG nova.compute.provider_tree [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 569.782125] env[63202]: DEBUG nova.network.neutron [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 570.219673] env[63202]: DEBUG nova.scheduler.client.report [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 570.285093] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Releasing lock "refresh_cache-65cd7042-daaf-4845-a0b8-5d2d0be02696" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 570.285537] env[63202]: DEBUG nova.compute.manager [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 570.286484] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 570.286609] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-52e3788d-2720-425c-93fb-ded26e31df67 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.297764] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16fd4324-d362-47b2-9f94-dd8438a25371 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.323037] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 65cd7042-daaf-4845-a0b8-5d2d0be02696 could not be found. [ 570.323037] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 570.323037] env[63202]: INFO nova.compute.manager [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] Took 0.04 seconds to destroy the instance on the hypervisor. [ 570.323200] env[63202]: DEBUG oslo.service.loopingcall [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 570.323954] env[63202]: DEBUG nova.compute.manager [-] [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 570.323954] env[63202]: DEBUG nova.network.neutron [-] [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 570.370026] env[63202]: DEBUG nova.network.neutron [-] [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 570.568072] env[63202]: DEBUG nova.compute.manager [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 570.599054] env[63202]: DEBUG nova.virt.hardware [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 570.599054] env[63202]: DEBUG nova.virt.hardware [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 570.599054] env[63202]: DEBUG nova.virt.hardware [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 570.599054] env[63202]: DEBUG nova.virt.hardware [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 570.599278] env[63202]: DEBUG nova.virt.hardware [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 570.599501] env[63202]: DEBUG nova.virt.hardware [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 570.600243] env[63202]: DEBUG nova.virt.hardware [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 570.601045] env[63202]: DEBUG nova.virt.hardware [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 570.601045] env[63202]: DEBUG nova.virt.hardware [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 570.601363] env[63202]: DEBUG nova.virt.hardware [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 570.601721] env[63202]: DEBUG nova.virt.hardware [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 570.603087] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc3938aa-0cc0-4666-9468-a6b815f3c58c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.611662] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21f32e25-4f94-4dcb-996f-533859ee468f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.626146] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Instance VIF info [] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 570.640229] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 570.640229] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f4c75660-f0b6-48c8-a7b0-21419dff9c8b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.654084] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Created folder: OpenStack in parent group-v4. [ 570.654084] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Creating folder: Project (a12be02b0b4e4bfe84aa8e064146582f). Parent ref: group-v294090. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 570.654084] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-816a7201-305c-4e87-b6b7-1c090bb47a3f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.661586] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Created folder: Project (a12be02b0b4e4bfe84aa8e064146582f) in parent group-v294090. [ 570.661777] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Creating folder: Instances. Parent ref: group-v294091. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 570.661988] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fdf87b65-8fec-4453-ba2d-3f22687cb278 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.670556] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Created folder: Instances in parent group-v294091. [ 570.670878] env[63202]: DEBUG oslo.service.loopingcall [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 570.671012] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 570.671180] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2599b3c4-5a4e-470c-95f9-b18ca6ccaa7a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.694309] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 570.694309] env[63202]: value = "task-1385270" [ 570.694309] env[63202]: _type = "Task" [ 570.694309] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 570.707490] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385270, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 570.724576] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.194s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 570.724576] env[63202]: DEBUG nova.compute.manager [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 570.873373] env[63202]: DEBUG nova.network.neutron [-] [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 571.080290] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] Acquiring lock "cc04e228-8b1a-4897-967e-131623434ac0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 571.080503] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] Lock "cc04e228-8b1a-4897-967e-131623434ac0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 571.206543] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385270, 'name': CreateVM_Task, 'duration_secs': 0.332022} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 571.206701] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 571.207803] env[63202]: DEBUG oslo_vmware.service [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8af2ea8a-0a1f-4ee3-a15f-f16f814b4408 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.215683] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 571.215873] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 571.216557] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 571.216854] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-734ec0a8-d773-4b85-acb0-2c88258f8531 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.221974] env[63202]: DEBUG oslo_vmware.api [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Waiting for the task: (returnval){ [ 571.221974] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52748e94-9024-fa9f-b572-4c5bb62a4f32" [ 571.221974] env[63202]: _type = "Task" [ 571.221974] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 571.229969] env[63202]: DEBUG oslo_vmware.api [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52748e94-9024-fa9f-b572-4c5bb62a4f32, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 571.231172] env[63202]: DEBUG nova.compute.utils [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 571.232660] env[63202]: DEBUG nova.compute.manager [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 571.232952] env[63202]: DEBUG nova.network.neutron [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 571.375610] env[63202]: INFO nova.compute.manager [-] [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] Took 1.05 seconds to deallocate network for instance. [ 571.378675] env[63202]: DEBUG nova.compute.claims [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 571.379461] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 571.379729] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 571.397586] env[63202]: DEBUG nova.policy [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '261d7ab4f5454bdaa7ca3fcf87bb25e8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c27239103d3e4acda1537ee7db22159e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 571.586942] env[63202]: DEBUG nova.compute.manager [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] [instance: cc04e228-8b1a-4897-967e-131623434ac0] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 571.733583] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 571.733826] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 571.734076] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 571.734213] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 571.734615] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 571.735151] env[63202]: DEBUG nova.compute.manager [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 571.740111] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bbb46c0b-73ca-4ac6-bd93-d04291e8e997 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.759709] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 571.759894] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 571.760713] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5271d528-de98-40d9-b17b-5cf4cdf845ca {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.774933] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-129f5752-94db-4044-a7c1-6730bba2c831 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.781429] env[63202]: DEBUG oslo_vmware.api [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Waiting for the task: (returnval){ [ 571.781429] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]523c3585-7132-329e-c65c-600689f4caa5" [ 571.781429] env[63202]: _type = "Task" [ 571.781429] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 571.795448] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Preparing fetch location {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 571.795698] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Creating directory with path [datastore1] vmware_temp/9ab11a24-f3ec-496c-8502-feae5b70f5c7/bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 571.795924] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d334193d-1ffd-4e4b-be7c-cf39feb10a0a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.819369] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Created directory with path [datastore1] vmware_temp/9ab11a24-f3ec-496c-8502-feae5b70f5c7/bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 571.819582] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Fetch image to [datastore1] vmware_temp/9ab11a24-f3ec-496c-8502-feae5b70f5c7/bb172aa4-a5a8-4395-9793-2416c30721cd/tmp-sparse.vmdk {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 571.820115] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Downloading image file data bb172aa4-a5a8-4395-9793-2416c30721cd to [datastore1] vmware_temp/9ab11a24-f3ec-496c-8502-feae5b70f5c7/bb172aa4-a5a8-4395-9793-2416c30721cd/tmp-sparse.vmdk on the data store datastore1 {{(pid=63202) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 571.821058] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9f56fa6-e2ba-4c88-a5eb-370cb92b914c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.829028] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70d5d5e9-9a10-465f-b2c2-9ae190449bf9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.839858] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5717e6d9-fb64-4b59-a05b-f6d49c008dca {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.880750] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a0daa86-ae0a-431a-83b5-f91320f21feb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.894336] env[63202]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-d3928f22-5c18-4cae-aa39-9da989a8e460 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.926562] env[63202]: DEBUG nova.virt.vmwareapi.images [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Downloading image file data bb172aa4-a5a8-4395-9793-2416c30721cd to the data store datastore1 {{(pid=63202) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 572.038028] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21722d5f-7441-4b67-b02d-826d6060f22d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.048187] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8901c521-ba9a-4f05-9c7a-255e00a3f496 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.084710] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] Acquiring lock "4a0241c3-8dea-4deb-80e1-66bb9cba10b9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.085489] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] Lock "4a0241c3-8dea-4deb-80e1-66bb9cba10b9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.087547] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c261db8-91a7-4565-8d73-015f51d018e3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.100157] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9b79567-ff14-49cc-8729-61fd71c70f49 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.117807] env[63202]: DEBUG nova.compute.provider_tree [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 572.131324] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.158369] env[63202]: DEBUG oslo_vmware.rw_handles [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/9ab11a24-f3ec-496c-8502-feae5b70f5c7/bb172aa4-a5a8-4395-9793-2416c30721cd/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=63202) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 572.591146] env[63202]: DEBUG nova.compute.manager [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 572.625359] env[63202]: DEBUG nova.scheduler.client.report [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 572.751245] env[63202]: DEBUG nova.compute.manager [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 572.792312] env[63202]: DEBUG nova.virt.hardware [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 572.792571] env[63202]: DEBUG nova.virt.hardware [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 572.792721] env[63202]: DEBUG nova.virt.hardware [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 572.792932] env[63202]: DEBUG nova.virt.hardware [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 572.793220] env[63202]: DEBUG nova.virt.hardware [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 572.793329] env[63202]: DEBUG nova.virt.hardware [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 572.793548] env[63202]: DEBUG nova.virt.hardware [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 572.793703] env[63202]: DEBUG nova.virt.hardware [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 572.793865] env[63202]: DEBUG nova.virt.hardware [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 572.794121] env[63202]: DEBUG nova.virt.hardware [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 572.794203] env[63202]: DEBUG nova.virt.hardware [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 572.797163] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70f58c7f-0dc7-496f-a2e7-25bcde2d37b4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.806317] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Acquiring lock "21d0b283-3bb4-4bda-8e62-c933c6de4927" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.806403] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Lock "21d0b283-3bb4-4bda-8e62-c933c6de4927" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.817584] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f446c50-d971-4c4a-b222-42afda1f04c8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.979581] env[63202]: DEBUG nova.network.neutron [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] Successfully created port: 67c7322d-5c8e-4e9c-832c-45dd23718e3a {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 572.994510] env[63202]: DEBUG oslo_vmware.rw_handles [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Completed reading data from the image iterator. {{(pid=63202) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 572.994510] env[63202]: DEBUG oslo_vmware.rw_handles [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Closing write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/9ab11a24-f3ec-496c-8502-feae5b70f5c7/bb172aa4-a5a8-4395-9793-2416c30721cd/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=63202) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 573.041008] env[63202]: DEBUG nova.virt.vmwareapi.images [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Downloaded image file data bb172aa4-a5a8-4395-9793-2416c30721cd to vmware_temp/9ab11a24-f3ec-496c-8502-feae5b70f5c7/bb172aa4-a5a8-4395-9793-2416c30721cd/tmp-sparse.vmdk on the data store datastore1 {{(pid=63202) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 573.042919] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Caching image {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 573.042990] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Copying Virtual Disk [datastore1] vmware_temp/9ab11a24-f3ec-496c-8502-feae5b70f5c7/bb172aa4-a5a8-4395-9793-2416c30721cd/tmp-sparse.vmdk to [datastore1] vmware_temp/9ab11a24-f3ec-496c-8502-feae5b70f5c7/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 573.043283] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-11c114f7-55fd-4f2a-8b3f-562aef359292 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.057807] env[63202]: DEBUG oslo_vmware.api [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Waiting for the task: (returnval){ [ 573.057807] env[63202]: value = "task-1385271" [ 573.057807] env[63202]: _type = "Task" [ 573.057807] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 573.069936] env[63202]: DEBUG oslo_vmware.api [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385271, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 573.128831] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.134839] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.755s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 573.135600] env[63202]: ERROR nova.compute.manager [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4cf12941-cfc8-4fbf-9cd0-12b2427b08ba, please check neutron logs for more information. [ 573.135600] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] Traceback (most recent call last): [ 573.135600] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 573.135600] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] self.driver.spawn(context, instance, image_meta, [ 573.135600] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 573.135600] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] self._vmops.spawn(context, instance, image_meta, injected_files, [ 573.135600] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 573.135600] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] vm_ref = self.build_virtual_machine(instance, [ 573.135600] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 573.135600] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] vif_infos = vmwarevif.get_vif_info(self._session, [ 573.135600] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 573.135910] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] for vif in network_info: [ 573.135910] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 573.135910] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] return self._sync_wrapper(fn, *args, **kwargs) [ 573.135910] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 573.135910] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] self.wait() [ 573.135910] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 573.135910] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] self[:] = self._gt.wait() [ 573.135910] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 573.135910] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] return self._exit_event.wait() [ 573.135910] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 573.135910] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] result = hub.switch() [ 573.135910] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 573.135910] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] return self.greenlet.switch() [ 573.139431] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 573.139431] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] result = function(*args, **kwargs) [ 573.139431] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 573.139431] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] return func(*args, **kwargs) [ 573.139431] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 573.139431] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] raise e [ 573.139431] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 573.139431] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] nwinfo = self.network_api.allocate_for_instance( [ 573.139431] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 573.139431] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] created_port_ids = self._update_ports_for_instance( [ 573.139431] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 573.139431] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] with excutils.save_and_reraise_exception(): [ 573.139431] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 573.139854] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] self.force_reraise() [ 573.139854] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 573.139854] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] raise self.value [ 573.139854] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 573.139854] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] updated_port = self._update_port( [ 573.139854] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 573.139854] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] _ensure_no_port_binding_failure(port) [ 573.139854] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 573.139854] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] raise exception.PortBindingFailed(port_id=port['id']) [ 573.139854] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] nova.exception.PortBindingFailed: Binding failed for port 4cf12941-cfc8-4fbf-9cd0-12b2427b08ba, please check neutron logs for more information. [ 573.139854] env[63202]: ERROR nova.compute.manager [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] [ 573.140169] env[63202]: DEBUG nova.compute.utils [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] Binding failed for port 4cf12941-cfc8-4fbf-9cd0-12b2427b08ba, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 573.140330] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.008s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.142814] env[63202]: INFO nova.compute.claims [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] [instance: cc04e228-8b1a-4897-967e-131623434ac0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 573.149158] env[63202]: DEBUG nova.compute.manager [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] Build of instance 65cd7042-daaf-4845-a0b8-5d2d0be02696 was re-scheduled: Binding failed for port 4cf12941-cfc8-4fbf-9cd0-12b2427b08ba, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 573.149631] env[63202]: DEBUG nova.compute.manager [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 573.149849] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Acquiring lock "refresh_cache-65cd7042-daaf-4845-a0b8-5d2d0be02696" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 573.149988] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Acquired lock "refresh_cache-65cd7042-daaf-4845-a0b8-5d2d0be02696" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 573.150155] env[63202]: DEBUG nova.network.neutron [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 573.318935] env[63202]: DEBUG nova.compute.manager [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] [instance: 21d0b283-3bb4-4bda-8e62-c933c6de4927] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 573.494856] env[63202]: DEBUG oslo_concurrency.lockutils [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Acquiring lock "4646821e-13ec-4413-aa82-1bbf6eb384ad" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.495101] env[63202]: DEBUG oslo_concurrency.lockutils [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Lock "4646821e-13ec-4413-aa82-1bbf6eb384ad" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.573885] env[63202]: DEBUG oslo_vmware.api [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385271, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 573.697720] env[63202]: DEBUG nova.network.neutron [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 573.781513] env[63202]: DEBUG nova.network.neutron [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 573.851390] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.998107] env[63202]: DEBUG nova.compute.manager [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 574.085589] env[63202]: DEBUG oslo_vmware.api [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385271, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.759299} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 574.085589] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Copied Virtual Disk [datastore1] vmware_temp/9ab11a24-f3ec-496c-8502-feae5b70f5c7/bb172aa4-a5a8-4395-9793-2416c30721cd/tmp-sparse.vmdk to [datastore1] vmware_temp/9ab11a24-f3ec-496c-8502-feae5b70f5c7/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 574.085589] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Deleting the datastore file [datastore1] vmware_temp/9ab11a24-f3ec-496c-8502-feae5b70f5c7/bb172aa4-a5a8-4395-9793-2416c30721cd/tmp-sparse.vmdk {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 574.085589] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-79714b05-cf27-45e0-9250-810ee9630b2c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.089326] env[63202]: DEBUG oslo_vmware.api [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Waiting for the task: (returnval){ [ 574.089326] env[63202]: value = "task-1385272" [ 574.089326] env[63202]: _type = "Task" [ 574.089326] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 574.099030] env[63202]: DEBUG oslo_vmware.api [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385272, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 574.281142] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee10698c-e63e-446f-84cd-c521160aa38c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.285779] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Releasing lock "refresh_cache-65cd7042-daaf-4845-a0b8-5d2d0be02696" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 574.286055] env[63202]: DEBUG nova.compute.manager [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 574.286234] env[63202]: DEBUG nova.compute.manager [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 574.286460] env[63202]: DEBUG nova.network.neutron [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 574.291035] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-823c27ec-0b87-4d42-ad21-4144bdfd5080 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.322160] env[63202]: DEBUG nova.network.neutron [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 574.324910] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21305693-fa46-4623-bb8a-68e0443ce150 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.331982] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38b2b5ac-f671-4ed2-a74b-6559b25e8772 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.349047] env[63202]: DEBUG nova.compute.provider_tree [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 574.523498] env[63202]: DEBUG oslo_concurrency.lockutils [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.599783] env[63202]: DEBUG oslo_vmware.api [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385272, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.104863} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 574.599783] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 574.600815] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Moving file from [datastore1] vmware_temp/9ab11a24-f3ec-496c-8502-feae5b70f5c7/bb172aa4-a5a8-4395-9793-2416c30721cd to [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd. {{(pid=63202) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 574.600815] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-dd116a1c-4df6-4803-b4e6-13a29cb19588 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.607644] env[63202]: DEBUG oslo_vmware.api [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Waiting for the task: (returnval){ [ 574.607644] env[63202]: value = "task-1385273" [ 574.607644] env[63202]: _type = "Task" [ 574.607644] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 574.625523] env[63202]: DEBUG oslo_vmware.api [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385273, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 574.827542] env[63202]: DEBUG nova.network.neutron [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.852659] env[63202]: DEBUG nova.scheduler.client.report [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 575.120733] env[63202]: DEBUG oslo_vmware.api [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385273, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.024547} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 575.120733] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] File moved {{(pid=63202) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 575.120733] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Cleaning up location [datastore1] vmware_temp/9ab11a24-f3ec-496c-8502-feae5b70f5c7 {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 575.123435] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Deleting the datastore file [datastore1] vmware_temp/9ab11a24-f3ec-496c-8502-feae5b70f5c7 {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 575.123567] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0bbc453e-fe37-4b44-89e2-f28b118071b5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.130271] env[63202]: DEBUG oslo_vmware.api [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Waiting for the task: (returnval){ [ 575.130271] env[63202]: value = "task-1385274" [ 575.130271] env[63202]: _type = "Task" [ 575.130271] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 575.138389] env[63202]: DEBUG oslo_vmware.api [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385274, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 575.330833] env[63202]: INFO nova.compute.manager [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 65cd7042-daaf-4845-a0b8-5d2d0be02696] Took 1.04 seconds to deallocate network for instance. [ 575.359035] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.219s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 575.359392] env[63202]: DEBUG nova.compute.manager [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] [instance: cc04e228-8b1a-4897-967e-131623434ac0] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 575.367260] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.236s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 575.367260] env[63202]: INFO nova.compute.claims [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 575.645350] env[63202]: DEBUG oslo_vmware.api [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385274, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.025133} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 575.645609] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 575.646339] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce98afb5-d035-4af2-92d8-7d1f6aa53c7e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.653908] env[63202]: DEBUG oslo_vmware.api [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Waiting for the task: (returnval){ [ 575.653908] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52ed6be8-d2cf-22f3-03bb-28b0c70e7933" [ 575.653908] env[63202]: _type = "Task" [ 575.653908] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 575.664416] env[63202]: DEBUG oslo_vmware.api [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52ed6be8-d2cf-22f3-03bb-28b0c70e7933, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 575.872468] env[63202]: DEBUG nova.compute.utils [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 575.879105] env[63202]: DEBUG nova.compute.manager [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] [instance: cc04e228-8b1a-4897-967e-131623434ac0] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 575.879425] env[63202]: DEBUG nova.network.neutron [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] [instance: cc04e228-8b1a-4897-967e-131623434ac0] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 576.058348] env[63202]: DEBUG nova.policy [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e7b02a295f4045ebac5e3a79b6bffe57', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd6fdb008429747bda964c5059595101c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 576.170116] env[63202]: DEBUG oslo_vmware.api [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52ed6be8-d2cf-22f3-03bb-28b0c70e7933, 'name': SearchDatastore_Task, 'duration_secs': 0.008969} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 576.170116] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 576.170482] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] 0ded82aa-c68d-4d11-9e4d-90ed9a19b708/0ded82aa-c68d-4d11-9e4d-90ed9a19b708.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 576.170593] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9038aef4-7359-41de-b6e2-a5e6374a78e3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.181662] env[63202]: DEBUG oslo_vmware.api [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Waiting for the task: (returnval){ [ 576.181662] env[63202]: value = "task-1385275" [ 576.181662] env[63202]: _type = "Task" [ 576.181662] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 576.191117] env[63202]: DEBUG oslo_vmware.api [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385275, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 576.381177] env[63202]: INFO nova.scheduler.client.report [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Deleted allocations for instance 65cd7042-daaf-4845-a0b8-5d2d0be02696 [ 576.389945] env[63202]: DEBUG nova.compute.manager [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] [instance: cc04e228-8b1a-4897-967e-131623434ac0] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 576.558713] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8424d901-6d44-4cf3-9055-c141567b9b48 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.570696] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73c1ee0b-e842-49d8-a610-6277ae380bfa {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.603497] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b4a37c9-9fe9-405c-bcf1-3f314eefbd48 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.611499] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ae80ea1-1445-4b13-907c-2ba20a3ebd8d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.627953] env[63202]: DEBUG nova.compute.provider_tree [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 576.696512] env[63202]: DEBUG oslo_vmware.api [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385275, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.495687} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 576.696512] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] 0ded82aa-c68d-4d11-9e4d-90ed9a19b708/0ded82aa-c68d-4d11-9e4d-90ed9a19b708.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 576.696512] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 576.696782] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8c1dec21-a337-4d1b-92dd-802eab0c8dec {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.710399] env[63202]: DEBUG oslo_vmware.api [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Waiting for the task: (returnval){ [ 576.710399] env[63202]: value = "task-1385276" [ 576.710399] env[63202]: _type = "Task" [ 576.710399] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 576.716261] env[63202]: DEBUG oslo_vmware.api [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385276, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 576.901571] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bd3b5d75-33ff-4804-bc28-409b12d4637e tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Lock "65cd7042-daaf-4845-a0b8-5d2d0be02696" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.937s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 577.134994] env[63202]: DEBUG nova.scheduler.client.report [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 577.221421] env[63202]: DEBUG oslo_vmware.api [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385276, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069828} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 577.221516] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 577.222679] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee87e989-58d7-4552-98f9-f3e7e1bbc2a7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.255134] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Reconfiguring VM instance instance-00000002 to attach disk [datastore1] 0ded82aa-c68d-4d11-9e4d-90ed9a19b708/0ded82aa-c68d-4d11-9e4d-90ed9a19b708.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 577.255446] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8aa71e90-14ac-4280-a4f8-83466330e520 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.277514] env[63202]: DEBUG oslo_vmware.api [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Waiting for the task: (returnval){ [ 577.277514] env[63202]: value = "task-1385277" [ 577.277514] env[63202]: _type = "Task" [ 577.277514] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 577.291595] env[63202]: DEBUG oslo_vmware.api [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385277, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 577.404084] env[63202]: DEBUG nova.compute.manager [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] [instance: cc04e228-8b1a-4897-967e-131623434ac0] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 577.451706] env[63202]: DEBUG nova.virt.hardware [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 577.451953] env[63202]: DEBUG nova.virt.hardware [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 577.452127] env[63202]: DEBUG nova.virt.hardware [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 577.452305] env[63202]: DEBUG nova.virt.hardware [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 577.452453] env[63202]: DEBUG nova.virt.hardware [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 577.452591] env[63202]: DEBUG nova.virt.hardware [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 577.452796] env[63202]: DEBUG nova.virt.hardware [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 577.452953] env[63202]: DEBUG nova.virt.hardware [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 577.454873] env[63202]: DEBUG nova.virt.hardware [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 577.454873] env[63202]: DEBUG nova.virt.hardware [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 577.454873] env[63202]: DEBUG nova.virt.hardware [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 577.454873] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d25abec4-6479-4f7b-8fad-3cd262a23293 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.472335] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ca5ee9d-1d07-4b81-a4f2-a2e5fe4b67ef {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.554685] env[63202]: DEBUG nova.network.neutron [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] [instance: cc04e228-8b1a-4897-967e-131623434ac0] Successfully created port: 9f2cc00a-a922-474a-b98d-202cd3fb5d72 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 577.642201] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.277s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 577.642694] env[63202]: DEBUG nova.compute.manager [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 577.646987] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.796s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 577.649457] env[63202]: INFO nova.compute.claims [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] [instance: 21d0b283-3bb4-4bda-8e62-c933c6de4927] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 577.789799] env[63202]: DEBUG oslo_vmware.api [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385277, 'name': ReconfigVM_Task, 'duration_secs': 0.296193} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 577.790078] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Reconfigured VM instance instance-00000002 to attach disk [datastore1] 0ded82aa-c68d-4d11-9e4d-90ed9a19b708/0ded82aa-c68d-4d11-9e4d-90ed9a19b708.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 577.790813] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-29beca64-f38f-46c1-802c-055cb1e6ecb2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.797619] env[63202]: DEBUG oslo_vmware.api [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Waiting for the task: (returnval){ [ 577.797619] env[63202]: value = "task-1385278" [ 577.797619] env[63202]: _type = "Task" [ 577.797619] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 577.807817] env[63202]: DEBUG oslo_vmware.api [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385278, 'name': Rename_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 578.156518] env[63202]: DEBUG nova.compute.utils [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 578.162660] env[63202]: DEBUG nova.compute.manager [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 578.162660] env[63202]: DEBUG nova.network.neutron [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 578.177284] env[63202]: ERROR nova.compute.manager [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 67c7322d-5c8e-4e9c-832c-45dd23718e3a, please check neutron logs for more information. [ 578.177284] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 578.177284] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 578.177284] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 578.177284] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 578.177284] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 578.177284] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 578.177284] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 578.177284] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 578.177284] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 578.177284] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 578.177284] env[63202]: ERROR nova.compute.manager raise self.value [ 578.177284] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 578.177284] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 578.177284] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 578.177284] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 578.177864] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 578.177864] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 578.177864] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 67c7322d-5c8e-4e9c-832c-45dd23718e3a, please check neutron logs for more information. [ 578.177864] env[63202]: ERROR nova.compute.manager [ 578.177864] env[63202]: Traceback (most recent call last): [ 578.177864] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 578.177864] env[63202]: listener.cb(fileno) [ 578.177864] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 578.177864] env[63202]: result = function(*args, **kwargs) [ 578.177864] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 578.177864] env[63202]: return func(*args, **kwargs) [ 578.177864] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 578.177864] env[63202]: raise e [ 578.177864] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 578.177864] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 578.177864] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 578.177864] env[63202]: created_port_ids = self._update_ports_for_instance( [ 578.177864] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 578.177864] env[63202]: with excutils.save_and_reraise_exception(): [ 578.177864] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 578.177864] env[63202]: self.force_reraise() [ 578.177864] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 578.177864] env[63202]: raise self.value [ 578.177864] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 578.177864] env[63202]: updated_port = self._update_port( [ 578.177864] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 578.177864] env[63202]: _ensure_no_port_binding_failure(port) [ 578.177864] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 578.177864] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 578.178804] env[63202]: nova.exception.PortBindingFailed: Binding failed for port 67c7322d-5c8e-4e9c-832c-45dd23718e3a, please check neutron logs for more information. [ 578.178804] env[63202]: Removing descriptor: 14 [ 578.178899] env[63202]: ERROR nova.compute.manager [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 67c7322d-5c8e-4e9c-832c-45dd23718e3a, please check neutron logs for more information. [ 578.178899] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] Traceback (most recent call last): [ 578.178899] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 578.178899] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] yield resources [ 578.178899] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 578.178899] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] self.driver.spawn(context, instance, image_meta, [ 578.178899] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 578.178899] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 578.178899] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 578.178899] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] vm_ref = self.build_virtual_machine(instance, [ 578.178899] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 578.180277] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] vif_infos = vmwarevif.get_vif_info(self._session, [ 578.180277] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 578.180277] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] for vif in network_info: [ 578.180277] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 578.180277] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] return self._sync_wrapper(fn, *args, **kwargs) [ 578.180277] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 578.180277] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] self.wait() [ 578.180277] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 578.180277] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] self[:] = self._gt.wait() [ 578.180277] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 578.180277] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] return self._exit_event.wait() [ 578.180277] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 578.180277] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] result = hub.switch() [ 578.182068] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 578.182068] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] return self.greenlet.switch() [ 578.182068] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 578.182068] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] result = function(*args, **kwargs) [ 578.182068] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 578.182068] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] return func(*args, **kwargs) [ 578.182068] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 578.182068] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] raise e [ 578.182068] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 578.182068] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] nwinfo = self.network_api.allocate_for_instance( [ 578.182068] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 578.182068] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] created_port_ids = self._update_ports_for_instance( [ 578.182068] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 578.182585] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] with excutils.save_and_reraise_exception(): [ 578.182585] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 578.182585] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] self.force_reraise() [ 578.182585] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 578.182585] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] raise self.value [ 578.182585] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 578.182585] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] updated_port = self._update_port( [ 578.182585] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 578.182585] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] _ensure_no_port_binding_failure(port) [ 578.182585] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 578.182585] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] raise exception.PortBindingFailed(port_id=port['id']) [ 578.182585] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] nova.exception.PortBindingFailed: Binding failed for port 67c7322d-5c8e-4e9c-832c-45dd23718e3a, please check neutron logs for more information. [ 578.182585] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] [ 578.182976] env[63202]: INFO nova.compute.manager [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] Terminating instance [ 578.184587] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] Acquiring lock "refresh_cache-d87b8f05-e3d6-4800-a635-af5b6f03373d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 578.184587] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] Acquired lock "refresh_cache-d87b8f05-e3d6-4800-a635-af5b6f03373d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 578.184587] env[63202]: DEBUG nova.network.neutron [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 578.263029] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 578.263029] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 578.263029] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Starting heal instance info cache {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 578.263029] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Rebuilding the list of instances to heal {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 578.309665] env[63202]: DEBUG oslo_vmware.api [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385278, 'name': Rename_Task, 'duration_secs': 0.133282} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 578.309665] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 578.309665] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9478c2e3-cecb-4b41-ae81-3159ed23e9b6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.315177] env[63202]: DEBUG oslo_vmware.api [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Waiting for the task: (returnval){ [ 578.315177] env[63202]: value = "task-1385279" [ 578.315177] env[63202]: _type = "Task" [ 578.315177] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 578.323583] env[63202]: DEBUG oslo_vmware.api [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385279, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 578.389388] env[63202]: DEBUG nova.policy [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2a3bd67c529547e9ac51411b3a358fa9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0a3c32b8c8624174a26b120f103f8683', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 578.663428] env[63202]: DEBUG nova.compute.manager [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 578.773794] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Skipping network cache update for instance because it is Building. {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 578.773794] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] Skipping network cache update for instance because it is Building. {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 578.773794] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: cc04e228-8b1a-4897-967e-131623434ac0] Skipping network cache update for instance because it is Building. {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 578.773794] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] Skipping network cache update for instance because it is Building. {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 578.773794] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 21d0b283-3bb4-4bda-8e62-c933c6de4927] Skipping network cache update for instance because it is Building. {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 578.774949] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Didn't find any instances for network info cache update. {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 578.775167] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 578.775405] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 578.775590] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 578.775770] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 578.775949] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 578.777821] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 578.779530] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63202) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 578.779530] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager.update_available_resource {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 578.805219] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f777a6c-40d8-4249-bacd-e81194dd2765 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.816222] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cc5c991-241f-446c-a214-3d95fc10ea4a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.856290] env[63202]: DEBUG nova.network.neutron [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 578.860809] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a850c81-12c0-4b43-a9d0-4e190b11fa52 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.869515] env[63202]: DEBUG oslo_vmware.api [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385279, 'name': PowerOnVM_Task, 'duration_secs': 0.438863} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 578.872539] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 578.873018] env[63202]: INFO nova.compute.manager [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Took 8.31 seconds to spawn the instance on the hypervisor. [ 578.873409] env[63202]: DEBUG nova.compute.manager [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 578.874269] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85ddb08e-425f-41f2-be36-22775b7cd2de {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.878786] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b822205-2966-4866-a38f-22cf402773fe {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.894228] env[63202]: DEBUG nova.compute.provider_tree [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 579.053422] env[63202]: DEBUG nova.network.neutron [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 579.283845] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.401839] env[63202]: DEBUG nova.scheduler.client.report [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 579.416653] env[63202]: INFO nova.compute.manager [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Took 13.09 seconds to build instance. [ 579.556180] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] Releasing lock "refresh_cache-d87b8f05-e3d6-4800-a635-af5b6f03373d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 579.557384] env[63202]: DEBUG nova.compute.manager [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 579.557384] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 579.557384] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6d88fbd4-7316-4f62-89ca-b74c0ac6936a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.567984] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f54af503-6bb2-4a2c-8cfc-3ddb99db9b1d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.597110] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d87b8f05-e3d6-4800-a635-af5b6f03373d could not be found. [ 579.597451] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 579.597640] env[63202]: INFO nova.compute.manager [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] Took 0.04 seconds to destroy the instance on the hypervisor. [ 579.597955] env[63202]: DEBUG oslo.service.loopingcall [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 579.599091] env[63202]: DEBUG nova.network.neutron [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] Successfully created port: 77702c21-b059-4fe9-80ca-9fe5548801d9 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 579.601616] env[63202]: DEBUG nova.compute.manager [-] [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 579.601704] env[63202]: DEBUG nova.network.neutron [-] [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 579.643534] env[63202]: DEBUG nova.network.neutron [-] [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 579.683190] env[63202]: DEBUG nova.compute.manager [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 579.717295] env[63202]: DEBUG nova.virt.hardware [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 579.717568] env[63202]: DEBUG nova.virt.hardware [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 579.717661] env[63202]: DEBUG nova.virt.hardware [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 579.718720] env[63202]: DEBUG nova.virt.hardware [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 579.718720] env[63202]: DEBUG nova.virt.hardware [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 579.720143] env[63202]: DEBUG nova.virt.hardware [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 579.720143] env[63202]: DEBUG nova.virt.hardware [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 579.720143] env[63202]: DEBUG nova.virt.hardware [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 579.720722] env[63202]: DEBUG nova.virt.hardware [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 579.720722] env[63202]: DEBUG nova.virt.hardware [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 579.721896] env[63202]: DEBUG nova.virt.hardware [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 579.725753] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a8a682f-6d34-4398-a485-893a06125c21 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.736117] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab2220e0-fe72-40d1-94f8-ac37f8e3ab75 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.913682] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.266s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 579.914314] env[63202]: DEBUG nova.compute.manager [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] [instance: 21d0b283-3bb4-4bda-8e62-c933c6de4927] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 579.917710] env[63202]: DEBUG oslo_concurrency.lockutils [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.394s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.920028] env[63202]: INFO nova.compute.claims [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 579.927022] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a8ad7e9e-777b-4a9a-829b-b223f2f96363 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Lock "0ded82aa-c68d-4d11-9e4d-90ed9a19b708" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.602s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 579.987178] env[63202]: DEBUG nova.compute.manager [req-a5cab33e-29a5-4796-92d7-be3d326035a5 req-b969b726-8e0c-4f14-b50d-866ce3e1c396 service nova] [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] Received event network-changed-67c7322d-5c8e-4e9c-832c-45dd23718e3a {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 579.987178] env[63202]: DEBUG nova.compute.manager [req-a5cab33e-29a5-4796-92d7-be3d326035a5 req-b969b726-8e0c-4f14-b50d-866ce3e1c396 service nova] [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] Refreshing instance network info cache due to event network-changed-67c7322d-5c8e-4e9c-832c-45dd23718e3a. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 579.987912] env[63202]: DEBUG oslo_concurrency.lockutils [req-a5cab33e-29a5-4796-92d7-be3d326035a5 req-b969b726-8e0c-4f14-b50d-866ce3e1c396 service nova] Acquiring lock "refresh_cache-d87b8f05-e3d6-4800-a635-af5b6f03373d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 579.987912] env[63202]: DEBUG oslo_concurrency.lockutils [req-a5cab33e-29a5-4796-92d7-be3d326035a5 req-b969b726-8e0c-4f14-b50d-866ce3e1c396 service nova] Acquired lock "refresh_cache-d87b8f05-e3d6-4800-a635-af5b6f03373d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 579.991042] env[63202]: DEBUG nova.network.neutron [req-a5cab33e-29a5-4796-92d7-be3d326035a5 req-b969b726-8e0c-4f14-b50d-866ce3e1c396 service nova] [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] Refreshing network info cache for port 67c7322d-5c8e-4e9c-832c-45dd23718e3a {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 580.144610] env[63202]: DEBUG nova.network.neutron [-] [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.425607] env[63202]: DEBUG nova.compute.utils [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 580.428941] env[63202]: DEBUG nova.compute.manager [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] [instance: 21d0b283-3bb4-4bda-8e62-c933c6de4927] Not allocating networking since 'none' was specified. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 580.532705] env[63202]: DEBUG nova.network.neutron [req-a5cab33e-29a5-4796-92d7-be3d326035a5 req-b969b726-8e0c-4f14-b50d-866ce3e1c396 service nova] [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 580.649610] env[63202]: INFO nova.compute.manager [-] [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] Took 1.05 seconds to deallocate network for instance. [ 580.652890] env[63202]: DEBUG nova.compute.claims [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 580.654138] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.666027] env[63202]: DEBUG nova.network.neutron [req-a5cab33e-29a5-4796-92d7-be3d326035a5 req-b969b726-8e0c-4f14-b50d-866ce3e1c396 service nova] [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.747945] env[63202]: DEBUG oslo_concurrency.lockutils [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] Acquiring lock "6e102794-152e-492e-b063-e1ebaffa82b6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.747945] env[63202]: DEBUG oslo_concurrency.lockutils [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] Lock "6e102794-152e-492e-b063-e1ebaffa82b6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.931335] env[63202]: DEBUG nova.compute.manager [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] [instance: 21d0b283-3bb4-4bda-8e62-c933c6de4927] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 581.172199] env[63202]: DEBUG oslo_concurrency.lockutils [req-a5cab33e-29a5-4796-92d7-be3d326035a5 req-b969b726-8e0c-4f14-b50d-866ce3e1c396 service nova] Releasing lock "refresh_cache-d87b8f05-e3d6-4800-a635-af5b6f03373d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 581.185308] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b0e8074-a49e-4267-bcb2-cfdda18333fb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.193404] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09bba665-eea5-47e7-b797-c684f8705c24 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.228285] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be4d5b75-32a6-4689-95e8-dd5901c85047 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.236930] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d55fde03-3fea-4448-8747-3f5d1bf2e3dc {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.254218] env[63202]: DEBUG nova.compute.manager [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 581.256409] env[63202]: DEBUG nova.compute.provider_tree [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 581.764624] env[63202]: DEBUG nova.scheduler.client.report [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 581.794801] env[63202]: DEBUG oslo_concurrency.lockutils [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.811896] env[63202]: ERROR nova.compute.manager [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9f2cc00a-a922-474a-b98d-202cd3fb5d72, please check neutron logs for more information. [ 581.811896] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 581.811896] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 581.811896] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 581.811896] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 581.811896] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 581.811896] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 581.811896] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 581.811896] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 581.811896] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 581.811896] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 581.811896] env[63202]: ERROR nova.compute.manager raise self.value [ 581.811896] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 581.811896] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 581.811896] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 581.811896] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 581.812709] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 581.812709] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 581.812709] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9f2cc00a-a922-474a-b98d-202cd3fb5d72, please check neutron logs for more information. [ 581.812709] env[63202]: ERROR nova.compute.manager [ 581.812709] env[63202]: Traceback (most recent call last): [ 581.812709] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 581.812709] env[63202]: listener.cb(fileno) [ 581.812709] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 581.812709] env[63202]: result = function(*args, **kwargs) [ 581.812709] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 581.812709] env[63202]: return func(*args, **kwargs) [ 581.812709] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 581.812709] env[63202]: raise e [ 581.812709] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 581.812709] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 581.812709] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 581.812709] env[63202]: created_port_ids = self._update_ports_for_instance( [ 581.812709] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 581.812709] env[63202]: with excutils.save_and_reraise_exception(): [ 581.812709] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 581.812709] env[63202]: self.force_reraise() [ 581.812709] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 581.812709] env[63202]: raise self.value [ 581.812709] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 581.812709] env[63202]: updated_port = self._update_port( [ 581.812709] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 581.812709] env[63202]: _ensure_no_port_binding_failure(port) [ 581.812709] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 581.812709] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 581.813472] env[63202]: nova.exception.PortBindingFailed: Binding failed for port 9f2cc00a-a922-474a-b98d-202cd3fb5d72, please check neutron logs for more information. [ 581.813472] env[63202]: Removing descriptor: 16 [ 581.813472] env[63202]: ERROR nova.compute.manager [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] [instance: cc04e228-8b1a-4897-967e-131623434ac0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9f2cc00a-a922-474a-b98d-202cd3fb5d72, please check neutron logs for more information. [ 581.813472] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] Traceback (most recent call last): [ 581.813472] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 581.813472] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] yield resources [ 581.813472] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 581.813472] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] self.driver.spawn(context, instance, image_meta, [ 581.813472] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 581.813472] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 581.813472] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 581.813472] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] vm_ref = self.build_virtual_machine(instance, [ 581.813779] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 581.813779] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] vif_infos = vmwarevif.get_vif_info(self._session, [ 581.813779] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 581.813779] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] for vif in network_info: [ 581.813779] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 581.813779] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] return self._sync_wrapper(fn, *args, **kwargs) [ 581.813779] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 581.813779] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] self.wait() [ 581.813779] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 581.813779] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] self[:] = self._gt.wait() [ 581.813779] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 581.813779] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] return self._exit_event.wait() [ 581.813779] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 581.814131] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] result = hub.switch() [ 581.814131] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 581.814131] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] return self.greenlet.switch() [ 581.814131] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 581.814131] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] result = function(*args, **kwargs) [ 581.814131] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 581.814131] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] return func(*args, **kwargs) [ 581.814131] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 581.814131] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] raise e [ 581.814131] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 581.814131] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] nwinfo = self.network_api.allocate_for_instance( [ 581.814131] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 581.814131] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] created_port_ids = self._update_ports_for_instance( [ 581.814538] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 581.814538] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] with excutils.save_and_reraise_exception(): [ 581.814538] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 581.814538] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] self.force_reraise() [ 581.814538] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 581.814538] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] raise self.value [ 581.814538] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 581.814538] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] updated_port = self._update_port( [ 581.814538] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 581.814538] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] _ensure_no_port_binding_failure(port) [ 581.814538] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 581.814538] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] raise exception.PortBindingFailed(port_id=port['id']) [ 581.814883] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] nova.exception.PortBindingFailed: Binding failed for port 9f2cc00a-a922-474a-b98d-202cd3fb5d72, please check neutron logs for more information. [ 581.814883] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] [ 581.814883] env[63202]: INFO nova.compute.manager [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] [instance: cc04e228-8b1a-4897-967e-131623434ac0] Terminating instance [ 581.817390] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] Acquiring lock "refresh_cache-cc04e228-8b1a-4897-967e-131623434ac0" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 581.817738] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] Acquired lock "refresh_cache-cc04e228-8b1a-4897-967e-131623434ac0" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 581.817738] env[63202]: DEBUG nova.network.neutron [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] [instance: cc04e228-8b1a-4897-967e-131623434ac0] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 581.942677] env[63202]: DEBUG nova.compute.manager [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] [instance: 21d0b283-3bb4-4bda-8e62-c933c6de4927] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 581.991492] env[63202]: DEBUG nova.virt.hardware [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 581.991726] env[63202]: DEBUG nova.virt.hardware [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 581.991880] env[63202]: DEBUG nova.virt.hardware [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 581.992707] env[63202]: DEBUG nova.virt.hardware [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 581.992707] env[63202]: DEBUG nova.virt.hardware [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 581.992707] env[63202]: DEBUG nova.virt.hardware [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 581.992707] env[63202]: DEBUG nova.virt.hardware [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 581.993906] env[63202]: DEBUG nova.virt.hardware [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 581.993906] env[63202]: DEBUG nova.virt.hardware [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 581.993906] env[63202]: DEBUG nova.virt.hardware [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 581.993906] env[63202]: DEBUG nova.virt.hardware [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 581.996031] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b1da632-8d2b-40f4-aa94-f46a18d995fd {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.008082] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4e48643-b942-421a-a16d-b9e713a3e8d3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.026025] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] [instance: 21d0b283-3bb4-4bda-8e62-c933c6de4927] Instance VIF info [] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 582.036628] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Creating folder: Project (eaea3f8cb4d24dfb82e53c1cb3aeb764). Parent ref: group-v294090. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 582.036628] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9b538fc8-6f27-4f1a-8e0a-a5a13867690b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.045632] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Created folder: Project (eaea3f8cb4d24dfb82e53c1cb3aeb764) in parent group-v294090. [ 582.045816] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Creating folder: Instances. Parent ref: group-v294094. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 582.046057] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0cfc5509-e03e-4870-a7e5-4a50ec1fe0ee {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.059727] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Created folder: Instances in parent group-v294094. [ 582.059727] env[63202]: DEBUG oslo.service.loopingcall [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 582.059878] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 21d0b283-3bb4-4bda-8e62-c933c6de4927] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 582.060064] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c8d76941-f1d9-4933-8352-0104f7712b2b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.078987] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 582.078987] env[63202]: value = "task-1385282" [ 582.078987] env[63202]: _type = "Task" [ 582.078987] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 582.086530] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385282, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 582.273937] env[63202]: DEBUG oslo_concurrency.lockutils [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.356s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 582.274532] env[63202]: DEBUG nova.compute.manager [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 582.280209] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 2.996s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.280209] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 582.280209] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63202) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 582.280209] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 1.627s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.284998] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ca0f0e3-2ce1-4868-b374-b4b760dbdade {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.294742] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a46e141e-1477-47e4-925e-204b2494b1f9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.312101] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14ff40be-b898-4417-8bda-102b0fc93c69 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.318977] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-858d4736-300b-497d-82b7-45452f30beae {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.358954] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181301MB free_disk=176GB free_vcpus=48 pci_devices=None {{(pid=63202) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 582.358954] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 582.360170] env[63202]: DEBUG nova.network.neutron [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] [instance: cc04e228-8b1a-4897-967e-131623434ac0] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 582.593420] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385282, 'name': CreateVM_Task, 'duration_secs': 0.310674} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 582.593420] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 21d0b283-3bb4-4bda-8e62-c933c6de4927] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 582.593420] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 582.593420] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 582.593716] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 582.593955] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5387463f-4a29-446f-b32a-8dd9d5981946 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.599025] env[63202]: DEBUG oslo_vmware.api [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Waiting for the task: (returnval){ [ 582.599025] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52eee549-efb7-cc18-854a-050bb7d47852" [ 582.599025] env[63202]: _type = "Task" [ 582.599025] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 582.603185] env[63202]: DEBUG nova.network.neutron [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] [instance: cc04e228-8b1a-4897-967e-131623434ac0] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 582.611882] env[63202]: DEBUG oslo_vmware.api [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52eee549-efb7-cc18-854a-050bb7d47852, 'name': SearchDatastore_Task, 'duration_secs': 0.00879} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 582.612489] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 582.612770] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] [instance: 21d0b283-3bb4-4bda-8e62-c933c6de4927] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 582.613063] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 582.613262] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 582.613489] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 582.613796] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-289972a8-1585-4496-8cd4-339476696ca7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.621959] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 582.622361] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 582.623240] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a588855f-796e-41df-b84d-7b2c55b7e5fc {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.628497] env[63202]: DEBUG oslo_vmware.api [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Waiting for the task: (returnval){ [ 582.628497] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52153d2f-b941-672c-3ac5-667ade220228" [ 582.628497] env[63202]: _type = "Task" [ 582.628497] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 582.639849] env[63202]: DEBUG oslo_vmware.api [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52153d2f-b941-672c-3ac5-667ade220228, 'name': SearchDatastore_Task, 'duration_secs': 0.007637} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 582.640663] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d799cb54-e52e-481b-9f35-3cb23b566e4f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.645825] env[63202]: DEBUG oslo_vmware.api [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Waiting for the task: (returnval){ [ 582.645825] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]523163b1-82d2-156c-3e19-4f11a24deb37" [ 582.645825] env[63202]: _type = "Task" [ 582.645825] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 582.653930] env[63202]: DEBUG oslo_vmware.api [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]523163b1-82d2-156c-3e19-4f11a24deb37, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 582.786753] env[63202]: DEBUG nova.compute.utils [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 582.793019] env[63202]: DEBUG nova.compute.manager [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 582.793019] env[63202]: DEBUG nova.network.neutron [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 582.918065] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d24bfe7c-4436-45af-97a9-5be267b6bf38 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.925820] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66ee45b3-eaf3-4138-b3e1-7d5542f1cea5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.957090] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-967651f5-8cb8-4e22-9488-34988de67837 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.964876] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c498dd3-c546-4a74-978c-bd21b7eb7b3b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.981358] env[63202]: DEBUG nova.compute.provider_tree [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 582.984092] env[63202]: DEBUG nova.policy [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '46863ba14f1c4829a24eca823e763f2f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b028e34225744668807aae95712a8d41', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 582.988064] env[63202]: ERROR nova.compute.manager [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 77702c21-b059-4fe9-80ca-9fe5548801d9, please check neutron logs for more information. [ 582.988064] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 582.988064] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 582.988064] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 582.988064] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 582.988064] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 582.988064] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 582.988064] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 582.988064] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.988064] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 582.988064] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.988064] env[63202]: ERROR nova.compute.manager raise self.value [ 582.988064] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 582.988064] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 582.988064] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.988064] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 582.988435] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.988435] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 582.988435] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 77702c21-b059-4fe9-80ca-9fe5548801d9, please check neutron logs for more information. [ 582.988435] env[63202]: ERROR nova.compute.manager [ 582.988435] env[63202]: Traceback (most recent call last): [ 582.988435] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 582.988435] env[63202]: listener.cb(fileno) [ 582.988435] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 582.988435] env[63202]: result = function(*args, **kwargs) [ 582.988435] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 582.988435] env[63202]: return func(*args, **kwargs) [ 582.988435] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 582.988435] env[63202]: raise e [ 582.988435] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 582.988435] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 582.988435] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 582.988435] env[63202]: created_port_ids = self._update_ports_for_instance( [ 582.988435] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 582.988435] env[63202]: with excutils.save_and_reraise_exception(): [ 582.988435] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.988435] env[63202]: self.force_reraise() [ 582.988435] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.988435] env[63202]: raise self.value [ 582.988435] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 582.988435] env[63202]: updated_port = self._update_port( [ 582.988435] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.988435] env[63202]: _ensure_no_port_binding_failure(port) [ 582.988435] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.988435] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 582.989103] env[63202]: nova.exception.PortBindingFailed: Binding failed for port 77702c21-b059-4fe9-80ca-9fe5548801d9, please check neutron logs for more information. [ 582.989103] env[63202]: Removing descriptor: 17 [ 582.989103] env[63202]: ERROR nova.compute.manager [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 77702c21-b059-4fe9-80ca-9fe5548801d9, please check neutron logs for more information. [ 582.989103] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] Traceback (most recent call last): [ 582.989103] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 582.989103] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] yield resources [ 582.989103] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 582.989103] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] self.driver.spawn(context, instance, image_meta, [ 582.989103] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 582.989103] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 582.989103] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 582.989103] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] vm_ref = self.build_virtual_machine(instance, [ 582.989592] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 582.989592] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] vif_infos = vmwarevif.get_vif_info(self._session, [ 582.989592] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 582.989592] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] for vif in network_info: [ 582.989592] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 582.989592] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] return self._sync_wrapper(fn, *args, **kwargs) [ 582.989592] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 582.989592] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] self.wait() [ 582.989592] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 582.989592] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] self[:] = self._gt.wait() [ 582.989592] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 582.989592] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] return self._exit_event.wait() [ 582.989592] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 582.989910] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] result = hub.switch() [ 582.989910] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 582.989910] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] return self.greenlet.switch() [ 582.989910] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 582.989910] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] result = function(*args, **kwargs) [ 582.989910] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 582.989910] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] return func(*args, **kwargs) [ 582.989910] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 582.989910] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] raise e [ 582.989910] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 582.989910] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] nwinfo = self.network_api.allocate_for_instance( [ 582.989910] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 582.989910] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] created_port_ids = self._update_ports_for_instance( [ 582.990244] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 582.990244] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] with excutils.save_and_reraise_exception(): [ 582.990244] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.990244] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] self.force_reraise() [ 582.990244] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.990244] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] raise self.value [ 582.990244] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 582.990244] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] updated_port = self._update_port( [ 582.990244] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.990244] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] _ensure_no_port_binding_failure(port) [ 582.990244] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.990244] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] raise exception.PortBindingFailed(port_id=port['id']) [ 582.991470] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] nova.exception.PortBindingFailed: Binding failed for port 77702c21-b059-4fe9-80ca-9fe5548801d9, please check neutron logs for more information. [ 582.991470] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] [ 582.991470] env[63202]: INFO nova.compute.manager [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] Terminating instance [ 582.991470] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] Acquiring lock "refresh_cache-4a0241c3-8dea-4deb-80e1-66bb9cba10b9" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 582.991470] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] Acquired lock "refresh_cache-4a0241c3-8dea-4deb-80e1-66bb9cba10b9" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 582.991929] env[63202]: DEBUG nova.network.neutron [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 583.107800] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] Releasing lock "refresh_cache-cc04e228-8b1a-4897-967e-131623434ac0" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 583.110032] env[63202]: DEBUG nova.compute.manager [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] [instance: cc04e228-8b1a-4897-967e-131623434ac0] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 583.110032] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] [instance: cc04e228-8b1a-4897-967e-131623434ac0] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 583.110032] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ae3d6fe8-2ccc-412c-aa74-a6c817c8c49d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.118902] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac08af84-a7fd-40e3-8666-d26f7542ed41 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.146527] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] [instance: cc04e228-8b1a-4897-967e-131623434ac0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance cc04e228-8b1a-4897-967e-131623434ac0 could not be found. [ 583.146747] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] [instance: cc04e228-8b1a-4897-967e-131623434ac0] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 583.146997] env[63202]: INFO nova.compute.manager [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] [instance: cc04e228-8b1a-4897-967e-131623434ac0] Took 0.04 seconds to destroy the instance on the hypervisor. [ 583.147261] env[63202]: DEBUG oslo.service.loopingcall [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 583.147572] env[63202]: DEBUG nova.compute.manager [-] [instance: cc04e228-8b1a-4897-967e-131623434ac0] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 583.147764] env[63202]: DEBUG nova.network.neutron [-] [instance: cc04e228-8b1a-4897-967e-131623434ac0] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 583.159814] env[63202]: DEBUG oslo_vmware.api [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]523163b1-82d2-156c-3e19-4f11a24deb37, 'name': SearchDatastore_Task, 'duration_secs': 0.007615} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 583.160173] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 583.160464] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] 21d0b283-3bb4-4bda-8e62-c933c6de4927/21d0b283-3bb4-4bda-8e62-c933c6de4927.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 583.160812] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dabd1ceb-2281-40ae-8097-fc5bc5e260ed {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.168208] env[63202]: DEBUG oslo_vmware.api [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Waiting for the task: (returnval){ [ 583.168208] env[63202]: value = "task-1385283" [ 583.168208] env[63202]: _type = "Task" [ 583.168208] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 583.176695] env[63202]: DEBUG oslo_vmware.api [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Task: {'id': task-1385283, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 583.197015] env[63202]: DEBUG nova.network.neutron [-] [instance: cc04e228-8b1a-4897-967e-131623434ac0] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 583.282750] env[63202]: DEBUG nova.compute.manager [req-3b0db8e1-f2d7-4578-b559-23f2744a02b2 req-039e6d4d-3dad-4cb3-85d8-6c4d01aed078 service nova] [instance: cc04e228-8b1a-4897-967e-131623434ac0] Received event network-changed-9f2cc00a-a922-474a-b98d-202cd3fb5d72 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 583.282979] env[63202]: DEBUG nova.compute.manager [req-3b0db8e1-f2d7-4578-b559-23f2744a02b2 req-039e6d4d-3dad-4cb3-85d8-6c4d01aed078 service nova] [instance: cc04e228-8b1a-4897-967e-131623434ac0] Refreshing instance network info cache due to event network-changed-9f2cc00a-a922-474a-b98d-202cd3fb5d72. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 583.283204] env[63202]: DEBUG oslo_concurrency.lockutils [req-3b0db8e1-f2d7-4578-b559-23f2744a02b2 req-039e6d4d-3dad-4cb3-85d8-6c4d01aed078 service nova] Acquiring lock "refresh_cache-cc04e228-8b1a-4897-967e-131623434ac0" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 583.283348] env[63202]: DEBUG oslo_concurrency.lockutils [req-3b0db8e1-f2d7-4578-b559-23f2744a02b2 req-039e6d4d-3dad-4cb3-85d8-6c4d01aed078 service nova] Acquired lock "refresh_cache-cc04e228-8b1a-4897-967e-131623434ac0" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 583.283506] env[63202]: DEBUG nova.network.neutron [req-3b0db8e1-f2d7-4578-b559-23f2744a02b2 req-039e6d4d-3dad-4cb3-85d8-6c4d01aed078 service nova] [instance: cc04e228-8b1a-4897-967e-131623434ac0] Refreshing network info cache for port 9f2cc00a-a922-474a-b98d-202cd3fb5d72 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 583.296922] env[63202]: DEBUG nova.compute.manager [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 583.486877] env[63202]: DEBUG nova.scheduler.client.report [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 583.561046] env[63202]: DEBUG nova.network.neutron [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 583.679458] env[63202]: DEBUG oslo_vmware.api [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Task: {'id': task-1385283, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.47395} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 583.679713] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] 21d0b283-3bb4-4bda-8e62-c933c6de4927/21d0b283-3bb4-4bda-8e62-c933c6de4927.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 583.679919] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] [instance: 21d0b283-3bb4-4bda-8e62-c933c6de4927] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 583.680183] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a60df62d-9b0f-4e20-8c25-c5ee66603584 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.687834] env[63202]: DEBUG oslo_vmware.api [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Waiting for the task: (returnval){ [ 583.687834] env[63202]: value = "task-1385284" [ 583.687834] env[63202]: _type = "Task" [ 583.687834] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 583.696563] env[63202]: DEBUG oslo_vmware.api [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Task: {'id': task-1385284, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 583.699038] env[63202]: DEBUG nova.network.neutron [-] [instance: cc04e228-8b1a-4897-967e-131623434ac0] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 583.829833] env[63202]: DEBUG nova.network.neutron [req-3b0db8e1-f2d7-4578-b559-23f2744a02b2 req-039e6d4d-3dad-4cb3-85d8-6c4d01aed078 service nova] [instance: cc04e228-8b1a-4897-967e-131623434ac0] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 583.885966] env[63202]: DEBUG nova.network.neutron [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 583.995791] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.716s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 583.996413] env[63202]: ERROR nova.compute.manager [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 67c7322d-5c8e-4e9c-832c-45dd23718e3a, please check neutron logs for more information. [ 583.996413] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] Traceback (most recent call last): [ 583.996413] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 583.996413] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] self.driver.spawn(context, instance, image_meta, [ 583.996413] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 583.996413] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 583.996413] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 583.996413] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] vm_ref = self.build_virtual_machine(instance, [ 583.996413] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 583.996413] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] vif_infos = vmwarevif.get_vif_info(self._session, [ 583.996413] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 583.997077] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] for vif in network_info: [ 583.997077] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 583.997077] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] return self._sync_wrapper(fn, *args, **kwargs) [ 583.997077] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 583.997077] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] self.wait() [ 583.997077] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 583.997077] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] self[:] = self._gt.wait() [ 583.997077] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 583.997077] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] return self._exit_event.wait() [ 583.997077] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 583.997077] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] result = hub.switch() [ 583.997077] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 583.997077] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] return self.greenlet.switch() [ 583.997467] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 583.997467] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] result = function(*args, **kwargs) [ 583.997467] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 583.997467] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] return func(*args, **kwargs) [ 583.997467] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 583.997467] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] raise e [ 583.997467] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 583.997467] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] nwinfo = self.network_api.allocate_for_instance( [ 583.997467] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 583.997467] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] created_port_ids = self._update_ports_for_instance( [ 583.997467] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 583.997467] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] with excutils.save_and_reraise_exception(): [ 583.997467] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 583.997874] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] self.force_reraise() [ 583.997874] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 583.997874] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] raise self.value [ 583.997874] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 583.997874] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] updated_port = self._update_port( [ 583.997874] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 583.997874] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] _ensure_no_port_binding_failure(port) [ 583.997874] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 583.997874] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] raise exception.PortBindingFailed(port_id=port['id']) [ 583.997874] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] nova.exception.PortBindingFailed: Binding failed for port 67c7322d-5c8e-4e9c-832c-45dd23718e3a, please check neutron logs for more information. [ 583.997874] env[63202]: ERROR nova.compute.manager [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] [ 583.998199] env[63202]: DEBUG nova.compute.utils [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] Binding failed for port 67c7322d-5c8e-4e9c-832c-45dd23718e3a, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 584.001286] env[63202]: DEBUG oslo_concurrency.lockutils [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.205s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 584.003215] env[63202]: INFO nova.compute.claims [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 584.007729] env[63202]: DEBUG nova.compute.manager [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] Build of instance d87b8f05-e3d6-4800-a635-af5b6f03373d was re-scheduled: Binding failed for port 67c7322d-5c8e-4e9c-832c-45dd23718e3a, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 584.008152] env[63202]: DEBUG nova.compute.manager [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 584.008371] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] Acquiring lock "refresh_cache-d87b8f05-e3d6-4800-a635-af5b6f03373d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 584.008504] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] Acquired lock "refresh_cache-d87b8f05-e3d6-4800-a635-af5b6f03373d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 584.008698] env[63202]: DEBUG nova.network.neutron [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 584.145645] env[63202]: DEBUG nova.network.neutron [req-3b0db8e1-f2d7-4578-b559-23f2744a02b2 req-039e6d4d-3dad-4cb3-85d8-6c4d01aed078 service nova] [instance: cc04e228-8b1a-4897-967e-131623434ac0] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 584.188530] env[63202]: DEBUG nova.network.neutron [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] Successfully created port: f8672ac4-0bf6-410d-af1e-af7434600f7d {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 584.202015] env[63202]: INFO nova.compute.manager [-] [instance: cc04e228-8b1a-4897-967e-131623434ac0] Took 1.05 seconds to deallocate network for instance. [ 584.205869] env[63202]: DEBUG nova.compute.claims [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] [instance: cc04e228-8b1a-4897-967e-131623434ac0] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 584.205869] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 584.208898] env[63202]: DEBUG oslo_vmware.api [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Task: {'id': task-1385284, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.262916} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 584.209142] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] [instance: 21d0b283-3bb4-4bda-8e62-c933c6de4927] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 584.210216] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f19723c6-9285-48e6-9c0e-da0d8b2ad2a9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.232498] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] [instance: 21d0b283-3bb4-4bda-8e62-c933c6de4927] Reconfiguring VM instance instance-00000006 to attach disk [datastore1] 21d0b283-3bb4-4bda-8e62-c933c6de4927/21d0b283-3bb4-4bda-8e62-c933c6de4927.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 584.234278] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7bf684ce-3f35-4a0f-a23c-58811fbee715 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.258164] env[63202]: DEBUG oslo_vmware.api [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Waiting for the task: (returnval){ [ 584.258164] env[63202]: value = "task-1385285" [ 584.258164] env[63202]: _type = "Task" [ 584.258164] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 584.270496] env[63202]: DEBUG oslo_vmware.api [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Task: {'id': task-1385285, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 584.306414] env[63202]: DEBUG nova.compute.manager [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 584.346294] env[63202]: DEBUG nova.virt.hardware [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 584.348211] env[63202]: DEBUG nova.virt.hardware [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 584.348211] env[63202]: DEBUG nova.virt.hardware [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 584.348211] env[63202]: DEBUG nova.virt.hardware [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 584.348211] env[63202]: DEBUG nova.virt.hardware [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 584.348211] env[63202]: DEBUG nova.virt.hardware [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 584.348555] env[63202]: DEBUG nova.virt.hardware [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 584.348555] env[63202]: DEBUG nova.virt.hardware [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 584.348555] env[63202]: DEBUG nova.virt.hardware [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 584.348555] env[63202]: DEBUG nova.virt.hardware [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 584.349290] env[63202]: DEBUG nova.virt.hardware [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 584.350406] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee4e8089-b46f-44f5-bd2d-69495bb2ae96 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.360750] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25e7b142-e67b-4f19-8563-5565c2bc6409 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.391389] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] Releasing lock "refresh_cache-4a0241c3-8dea-4deb-80e1-66bb9cba10b9" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 584.394639] env[63202]: DEBUG nova.compute.manager [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 584.394639] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 584.394639] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f6f74d40-35e2-4dd0-b6c3-1be762497287 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.403843] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6270597-749b-463e-9b43-ba0ba551632e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.435998] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4a0241c3-8dea-4deb-80e1-66bb9cba10b9 could not be found. [ 584.436661] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 584.436661] env[63202]: INFO nova.compute.manager [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] Took 0.04 seconds to destroy the instance on the hypervisor. [ 584.436661] env[63202]: DEBUG oslo.service.loopingcall [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 584.436903] env[63202]: DEBUG nova.compute.manager [-] [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 584.437260] env[63202]: DEBUG nova.network.neutron [-] [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 584.556378] env[63202]: DEBUG nova.network.neutron [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 584.584850] env[63202]: DEBUG nova.network.neutron [-] [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 584.652535] env[63202]: DEBUG oslo_concurrency.lockutils [req-3b0db8e1-f2d7-4578-b559-23f2744a02b2 req-039e6d4d-3dad-4cb3-85d8-6c4d01aed078 service nova] Releasing lock "refresh_cache-cc04e228-8b1a-4897-967e-131623434ac0" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 584.719092] env[63202]: DEBUG nova.network.neutron [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 584.753551] env[63202]: INFO nova.compute.manager [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Rebuilding instance [ 584.775857] env[63202]: DEBUG oslo_vmware.api [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Task: {'id': task-1385285, 'name': ReconfigVM_Task, 'duration_secs': 0.330523} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 584.778996] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] [instance: 21d0b283-3bb4-4bda-8e62-c933c6de4927] Reconfigured VM instance instance-00000006 to attach disk [datastore1] 21d0b283-3bb4-4bda-8e62-c933c6de4927/21d0b283-3bb4-4bda-8e62-c933c6de4927.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 584.779712] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0d80abd2-375a-457f-8b7c-a4cb2c291b11 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.790165] env[63202]: DEBUG oslo_vmware.api [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Waiting for the task: (returnval){ [ 584.790165] env[63202]: value = "task-1385286" [ 584.790165] env[63202]: _type = "Task" [ 584.790165] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 584.802236] env[63202]: DEBUG oslo_vmware.api [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Task: {'id': task-1385286, 'name': Rename_Task} progress is 6%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 584.833460] env[63202]: DEBUG nova.compute.manager [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 584.834357] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f636a709-4359-456f-b625-c4bac7f64e1a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.091598] env[63202]: DEBUG nova.network.neutron [-] [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.219946] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d098f71d-fd76-4721-934c-f5154e8a000a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.225475] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] Releasing lock "refresh_cache-d87b8f05-e3d6-4800-a635-af5b6f03373d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 585.225647] env[63202]: DEBUG nova.compute.manager [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 585.225798] env[63202]: DEBUG nova.compute.manager [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 585.225958] env[63202]: DEBUG nova.network.neutron [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 585.239403] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07ab16af-700b-49f8-8c90-ab8de218082d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.274466] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-335a5a97-e667-46fd-bf0f-765141ffcff5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.287955] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afbc3b16-f070-45cb-b271-ad495ccf469a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.305501] env[63202]: DEBUG oslo_concurrency.lockutils [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] Acquiring lock "90999eb1-8829-4f99-b3b9-8c70ee636cc5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.305722] env[63202]: DEBUG oslo_concurrency.lockutils [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] Lock "90999eb1-8829-4f99-b3b9-8c70ee636cc5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.316081] env[63202]: DEBUG nova.compute.provider_tree [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 585.321114] env[63202]: DEBUG nova.network.neutron [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 585.330494] env[63202]: DEBUG oslo_vmware.api [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Task: {'id': task-1385286, 'name': Rename_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 585.353787] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 585.354183] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-680f6c04-5d26-4453-b13e-f81a62835d94 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.363698] env[63202]: DEBUG oslo_vmware.api [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Waiting for the task: (returnval){ [ 585.363698] env[63202]: value = "task-1385287" [ 585.363698] env[63202]: _type = "Task" [ 585.363698] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 585.377559] env[63202]: DEBUG oslo_vmware.api [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385287, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 585.601625] env[63202]: INFO nova.compute.manager [-] [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] Took 1.16 seconds to deallocate network for instance. [ 585.607711] env[63202]: DEBUG nova.compute.claims [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 585.607711] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.807870] env[63202]: DEBUG oslo_vmware.api [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Task: {'id': task-1385286, 'name': Rename_Task, 'duration_secs': 0.880588} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 585.809480] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] [instance: 21d0b283-3bb4-4bda-8e62-c933c6de4927] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 585.809480] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ef8ed6d4-2b38-4008-a0bd-f1178eb2485c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.815217] env[63202]: DEBUG oslo_vmware.api [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Waiting for the task: (returnval){ [ 585.815217] env[63202]: value = "task-1385288" [ 585.815217] env[63202]: _type = "Task" [ 585.815217] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 585.824426] env[63202]: DEBUG nova.compute.manager [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 585.827912] env[63202]: DEBUG nova.scheduler.client.report [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 585.831184] env[63202]: DEBUG oslo_vmware.api [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Task: {'id': task-1385288, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 585.839908] env[63202]: DEBUG nova.network.neutron [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.874396] env[63202]: DEBUG oslo_vmware.api [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385287, 'name': PowerOffVM_Task, 'duration_secs': 0.156775} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 585.875201] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 585.875582] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 585.878045] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04a058ff-c64f-4ade-9760-228acba1851f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.886169] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 585.888351] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-499218fc-e19b-46f2-a880-3f7732acd0b6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.909608] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 585.910723] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Deleting contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 585.912019] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Deleting the datastore file [datastore1] 0ded82aa-c68d-4d11-9e4d-90ed9a19b708 {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 585.912019] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-30b0a41e-e8ca-4653-b587-da570d17d137 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.919571] env[63202]: DEBUG oslo_vmware.api [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Waiting for the task: (returnval){ [ 585.919571] env[63202]: value = "task-1385290" [ 585.919571] env[63202]: _type = "Task" [ 585.919571] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 585.929713] env[63202]: DEBUG oslo_vmware.api [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385290, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 586.329029] env[63202]: DEBUG oslo_vmware.api [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Task: {'id': task-1385288, 'name': PowerOnVM_Task, 'duration_secs': 0.459909} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 586.329212] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] [instance: 21d0b283-3bb4-4bda-8e62-c933c6de4927] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 586.329445] env[63202]: INFO nova.compute.manager [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] [instance: 21d0b283-3bb4-4bda-8e62-c933c6de4927] Took 4.39 seconds to spawn the instance on the hypervisor. [ 586.329612] env[63202]: DEBUG nova.compute.manager [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] [instance: 21d0b283-3bb4-4bda-8e62-c933c6de4927] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 586.330593] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6e24831-3a12-4540-bc27-d75215d2e4d2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.337089] env[63202]: DEBUG oslo_concurrency.lockutils [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.338s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 586.337541] env[63202]: DEBUG nova.compute.manager [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 586.343402] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 3.983s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 586.344388] env[63202]: INFO nova.compute.manager [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] [instance: d87b8f05-e3d6-4800-a635-af5b6f03373d] Took 1.12 seconds to deallocate network for instance. [ 586.369822] env[63202]: DEBUG oslo_concurrency.lockutils [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 586.437296] env[63202]: DEBUG oslo_vmware.api [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385290, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.112369} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 586.438094] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 586.438094] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Deleted contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 586.438094] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 586.855964] env[63202]: DEBUG nova.compute.utils [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 586.867404] env[63202]: DEBUG nova.compute.manager [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 586.867404] env[63202]: DEBUG nova.network.neutron [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 586.876673] env[63202]: INFO nova.compute.manager [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] [instance: 21d0b283-3bb4-4bda-8e62-c933c6de4927] Took 13.05 seconds to build instance. [ 587.224504] env[63202]: DEBUG nova.policy [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '219b392f66d841af988d641075a57382', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a07d44cdf93e4580b82fdc87d947aee6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 587.361675] env[63202]: DEBUG nova.compute.manager [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 587.381827] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bbc9bd11-081e-43a5-9701-629aedeb4f09 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Lock "21d0b283-3bb4-4bda-8e62-c933c6de4927" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.575s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 587.393233] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 0ded82aa-c68d-4d11-9e4d-90ed9a19b708 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 587.423950] env[63202]: INFO nova.scheduler.client.report [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] Deleted allocations for instance d87b8f05-e3d6-4800-a635-af5b6f03373d [ 587.483409] env[63202]: DEBUG nova.virt.hardware [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 587.483636] env[63202]: DEBUG nova.virt.hardware [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 587.483782] env[63202]: DEBUG nova.virt.hardware [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 587.483971] env[63202]: DEBUG nova.virt.hardware [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 587.486488] env[63202]: DEBUG nova.virt.hardware [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 587.486488] env[63202]: DEBUG nova.virt.hardware [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 587.486488] env[63202]: DEBUG nova.virt.hardware [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 587.486488] env[63202]: DEBUG nova.virt.hardware [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 587.486488] env[63202]: DEBUG nova.virt.hardware [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 587.486713] env[63202]: DEBUG nova.virt.hardware [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 587.486713] env[63202]: DEBUG nova.virt.hardware [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 587.486713] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74966e91-9ca0-48fc-90a7-dc935fbabc41 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.495121] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-480c9efb-1432-419a-9012-81da1eb705cb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.509088] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Instance VIF info [] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 587.514642] env[63202]: DEBUG oslo.service.loopingcall [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 587.514911] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 587.515130] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4a10fc4c-2f45-4ab9-bd33-0d5366e20490 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.531974] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 587.531974] env[63202]: value = "task-1385291" [ 587.531974] env[63202]: _type = "Task" [ 587.531974] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 587.539950] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385291, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 587.672311] env[63202]: DEBUG nova.compute.manager [req-14bd633f-0657-4d5a-937b-f47be54f7fb7 req-00d8fbbc-3328-4158-a15c-e629a26525cf service nova] [instance: cc04e228-8b1a-4897-967e-131623434ac0] Received event network-vif-deleted-9f2cc00a-a922-474a-b98d-202cd3fb5d72 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 587.672311] env[63202]: DEBUG nova.compute.manager [req-14bd633f-0657-4d5a-937b-f47be54f7fb7 req-00d8fbbc-3328-4158-a15c-e629a26525cf service nova] [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] Received event network-changed-77702c21-b059-4fe9-80ca-9fe5548801d9 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 587.672311] env[63202]: DEBUG nova.compute.manager [req-14bd633f-0657-4d5a-937b-f47be54f7fb7 req-00d8fbbc-3328-4158-a15c-e629a26525cf service nova] [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] Refreshing instance network info cache due to event network-changed-77702c21-b059-4fe9-80ca-9fe5548801d9. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 587.672311] env[63202]: DEBUG oslo_concurrency.lockutils [req-14bd633f-0657-4d5a-937b-f47be54f7fb7 req-00d8fbbc-3328-4158-a15c-e629a26525cf service nova] Acquiring lock "refresh_cache-4a0241c3-8dea-4deb-80e1-66bb9cba10b9" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 587.672311] env[63202]: DEBUG oslo_concurrency.lockutils [req-14bd633f-0657-4d5a-937b-f47be54f7fb7 req-00d8fbbc-3328-4158-a15c-e629a26525cf service nova] Acquired lock "refresh_cache-4a0241c3-8dea-4deb-80e1-66bb9cba10b9" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 587.672707] env[63202]: DEBUG nova.network.neutron [req-14bd633f-0657-4d5a-937b-f47be54f7fb7 req-00d8fbbc-3328-4158-a15c-e629a26525cf service nova] [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] Refreshing network info cache for port 77702c21-b059-4fe9-80ca-9fe5548801d9 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 587.897273] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance d87b8f05-e3d6-4800-a635-af5b6f03373d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 587.897590] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance cc04e228-8b1a-4897-967e-131623434ac0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 587.897590] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 4a0241c3-8dea-4deb-80e1-66bb9cba10b9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 587.897670] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 21d0b283-3bb4-4bda-8e62-c933c6de4927 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 587.897964] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 4646821e-13ec-4413-aa82-1bbf6eb384ad actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 587.897964] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 6e102794-152e-492e-b063-e1ebaffa82b6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 587.933212] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7acb3682-4c5d-447c-9ea7-77e43f694113 tempest-TenantUsagesTestJSON-1926285020 tempest-TenantUsagesTestJSON-1926285020-project-member] Lock "d87b8f05-e3d6-4800-a635-af5b6f03373d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.344s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 588.045482] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385291, 'name': CreateVM_Task, 'duration_secs': 0.308059} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 588.045482] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 588.046466] env[63202]: DEBUG oslo_vmware.service [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-557a14c3-b9a5-4712-9eee-f716e3eeb813 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.054503] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 588.054690] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 588.055173] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 588.055332] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5bbce788-2f88-4b65-a417-10df2b3f2927 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.061301] env[63202]: DEBUG oslo_vmware.api [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Waiting for the task: (returnval){ [ 588.061301] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]520b690d-c219-d4d2-8956-9a73db4b54f4" [ 588.061301] env[63202]: _type = "Task" [ 588.061301] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 588.072370] env[63202]: DEBUG oslo_vmware.api [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]520b690d-c219-d4d2-8956-9a73db4b54f4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 588.203325] env[63202]: DEBUG nova.network.neutron [req-14bd633f-0657-4d5a-937b-f47be54f7fb7 req-00d8fbbc-3328-4158-a15c-e629a26525cf service nova] [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 588.384742] env[63202]: DEBUG nova.compute.manager [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 588.403987] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 90999eb1-8829-4f99-b3b9-8c70ee636cc5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 588.404855] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=63202) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 588.404855] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=63202) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 588.424363] env[63202]: DEBUG nova.virt.hardware [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 588.424619] env[63202]: DEBUG nova.virt.hardware [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 588.424781] env[63202]: DEBUG nova.virt.hardware [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 588.424984] env[63202]: DEBUG nova.virt.hardware [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 588.428737] env[63202]: DEBUG nova.virt.hardware [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 588.428737] env[63202]: DEBUG nova.virt.hardware [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 588.429102] env[63202]: DEBUG nova.virt.hardware [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 588.429230] env[63202]: DEBUG nova.virt.hardware [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 588.429417] env[63202]: DEBUG nova.virt.hardware [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 588.429807] env[63202]: DEBUG nova.virt.hardware [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 588.430082] env[63202]: DEBUG nova.virt.hardware [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 588.431086] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04f3d9bd-1b4b-4a47-b7f1-b4ad156b5c80 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.447313] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a9c3e9e-d659-4d52-9a75-6bd167aee35e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.461077] env[63202]: DEBUG nova.network.neutron [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] Successfully created port: 5305bbd8-e698-4cb8-a36d-de3f0148ecac {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 588.475862] env[63202]: DEBUG nova.network.neutron [req-14bd633f-0657-4d5a-937b-f47be54f7fb7 req-00d8fbbc-3328-4158-a15c-e629a26525cf service nova] [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.556734] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae14a456-7c16-47b7-b7b7-533bea3e7813 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.577296] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e65341f6-f85a-4244-b03d-388bb119d52e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.581711] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 588.581889] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 588.582144] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 588.582322] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 588.582496] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 588.583265] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1a0baea4-3171-46e4-93dc-ef60cbb3c338 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.615409] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd0c1198-27c5-4c00-ae59-90d7fa55ca61 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.618112] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 588.618844] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 588.620760] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98807286-58ff-4523-a530-5752b5560e94 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.631136] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69c1748f-44b8-476a-b8cf-bb8f704a7573 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.635491] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d950778a-62a0-43db-9dd8-daaef40cf1e6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.651157] env[63202]: DEBUG nova.compute.provider_tree [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 588.651670] env[63202]: DEBUG oslo_vmware.api [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Waiting for the task: (returnval){ [ 588.651670] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52ced404-3fc3-1a8f-5b7b-1c934edc21f6" [ 588.651670] env[63202]: _type = "Task" [ 588.651670] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 588.664371] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Preparing fetch location {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 588.664763] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Creating directory with path [datastore2] vmware_temp/b4c81966-06b3-4ef8-acd4-8bd374d92f76/bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 588.664886] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3f0771d2-f8b0-43e5-a6ed-9e833fcc6bd3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.694627] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Created directory with path [datastore2] vmware_temp/b4c81966-06b3-4ef8-acd4-8bd374d92f76/bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 588.694627] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Fetch image to [datastore2] vmware_temp/b4c81966-06b3-4ef8-acd4-8bd374d92f76/bb172aa4-a5a8-4395-9793-2416c30721cd/tmp-sparse.vmdk {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 588.694767] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Downloading image file data bb172aa4-a5a8-4395-9793-2416c30721cd to [datastore2] vmware_temp/b4c81966-06b3-4ef8-acd4-8bd374d92f76/bb172aa4-a5a8-4395-9793-2416c30721cd/tmp-sparse.vmdk on the data store datastore2 {{(pid=63202) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 588.695658] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d032990a-7b9f-4304-bac6-c6cdfeb2b8ca {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.704311] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0aec1f5-8aa2-494d-9882-625d8ea62e94 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.714789] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70c66d59-b3c7-45f4-8bc2-b22748830d74 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.750961] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d5755cc-bd8b-4bac-9a9d-21130fecab89 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.758044] env[63202]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-c554c32b-b166-436e-8723-44158ff69675 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.791627] env[63202]: DEBUG nova.virt.vmwareapi.images [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Downloading image file data bb172aa4-a5a8-4395-9793-2416c30721cd to the data store datastore2 {{(pid=63202) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 588.875206] env[63202]: DEBUG oslo_vmware.rw_handles [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/b4c81966-06b3-4ef8-acd4-8bd374d92f76/bb172aa4-a5a8-4395-9793-2416c30721cd/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=63202) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 588.953483] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Acquiring lock "c538fef4-2245-4bd3-af03-f0105fceb1e3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 588.953746] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Lock "c538fef4-2245-4bd3-af03-f0105fceb1e3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.981083] env[63202]: DEBUG oslo_concurrency.lockutils [req-14bd633f-0657-4d5a-937b-f47be54f7fb7 req-00d8fbbc-3328-4158-a15c-e629a26525cf service nova] Releasing lock "refresh_cache-4a0241c3-8dea-4deb-80e1-66bb9cba10b9" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 588.982328] env[63202]: DEBUG nova.compute.manager [req-14bd633f-0657-4d5a-937b-f47be54f7fb7 req-00d8fbbc-3328-4158-a15c-e629a26525cf service nova] [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] Received event network-vif-deleted-77702c21-b059-4fe9-80ca-9fe5548801d9 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 589.131020] env[63202]: ERROR nova.compute.manager [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f8672ac4-0bf6-410d-af1e-af7434600f7d, please check neutron logs for more information. [ 589.131020] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 589.131020] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 589.131020] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 589.131020] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 589.131020] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 589.131020] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 589.131020] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 589.131020] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.131020] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 589.131020] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.131020] env[63202]: ERROR nova.compute.manager raise self.value [ 589.131020] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 589.131020] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 589.131020] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.131020] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 589.131614] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.131614] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 589.131614] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f8672ac4-0bf6-410d-af1e-af7434600f7d, please check neutron logs for more information. [ 589.131614] env[63202]: ERROR nova.compute.manager [ 589.131614] env[63202]: Traceback (most recent call last): [ 589.131614] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 589.131614] env[63202]: listener.cb(fileno) [ 589.131614] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 589.131614] env[63202]: result = function(*args, **kwargs) [ 589.131614] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 589.131614] env[63202]: return func(*args, **kwargs) [ 589.131614] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 589.131614] env[63202]: raise e [ 589.131614] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 589.131614] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 589.131614] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 589.131614] env[63202]: created_port_ids = self._update_ports_for_instance( [ 589.131614] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 589.131614] env[63202]: with excutils.save_and_reraise_exception(): [ 589.131614] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.131614] env[63202]: self.force_reraise() [ 589.131614] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.131614] env[63202]: raise self.value [ 589.131614] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 589.131614] env[63202]: updated_port = self._update_port( [ 589.131614] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.131614] env[63202]: _ensure_no_port_binding_failure(port) [ 589.131614] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.131614] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 589.132388] env[63202]: nova.exception.PortBindingFailed: Binding failed for port f8672ac4-0bf6-410d-af1e-af7434600f7d, please check neutron logs for more information. [ 589.132388] env[63202]: Removing descriptor: 16 [ 589.132388] env[63202]: ERROR nova.compute.manager [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f8672ac4-0bf6-410d-af1e-af7434600f7d, please check neutron logs for more information. [ 589.132388] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] Traceback (most recent call last): [ 589.132388] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 589.132388] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] yield resources [ 589.132388] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 589.132388] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] self.driver.spawn(context, instance, image_meta, [ 589.132388] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 589.132388] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] self._vmops.spawn(context, instance, image_meta, injected_files, [ 589.132388] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 589.132388] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] vm_ref = self.build_virtual_machine(instance, [ 589.132712] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 589.132712] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] vif_infos = vmwarevif.get_vif_info(self._session, [ 589.132712] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 589.132712] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] for vif in network_info: [ 589.132712] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 589.132712] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] return self._sync_wrapper(fn, *args, **kwargs) [ 589.132712] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 589.132712] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] self.wait() [ 589.132712] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 589.132712] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] self[:] = self._gt.wait() [ 589.132712] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 589.132712] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] return self._exit_event.wait() [ 589.132712] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 589.133081] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] result = hub.switch() [ 589.133081] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 589.133081] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] return self.greenlet.switch() [ 589.133081] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 589.133081] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] result = function(*args, **kwargs) [ 589.133081] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 589.133081] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] return func(*args, **kwargs) [ 589.133081] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 589.133081] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] raise e [ 589.133081] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 589.133081] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] nwinfo = self.network_api.allocate_for_instance( [ 589.133081] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 589.133081] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] created_port_ids = self._update_ports_for_instance( [ 589.137176] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 589.137176] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] with excutils.save_and_reraise_exception(): [ 589.137176] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.137176] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] self.force_reraise() [ 589.137176] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.137176] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] raise self.value [ 589.137176] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 589.137176] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] updated_port = self._update_port( [ 589.137176] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.137176] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] _ensure_no_port_binding_failure(port) [ 589.137176] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.137176] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] raise exception.PortBindingFailed(port_id=port['id']) [ 589.137783] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] nova.exception.PortBindingFailed: Binding failed for port f8672ac4-0bf6-410d-af1e-af7434600f7d, please check neutron logs for more information. [ 589.137783] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] [ 589.137783] env[63202]: INFO nova.compute.manager [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] Terminating instance [ 589.137783] env[63202]: DEBUG oslo_concurrency.lockutils [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Acquiring lock "refresh_cache-4646821e-13ec-4413-aa82-1bbf6eb384ad" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 589.137783] env[63202]: DEBUG oslo_concurrency.lockutils [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Acquired lock "refresh_cache-4646821e-13ec-4413-aa82-1bbf6eb384ad" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 589.137783] env[63202]: DEBUG nova.network.neutron [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 589.155521] env[63202]: DEBUG nova.scheduler.client.report [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 589.458240] env[63202]: DEBUG nova.compute.manager [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 589.606079] env[63202]: DEBUG oslo_vmware.rw_handles [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Completed reading data from the image iterator. {{(pid=63202) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 589.606769] env[63202]: DEBUG oslo_vmware.rw_handles [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Closing write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/b4c81966-06b3-4ef8-acd4-8bd374d92f76/bb172aa4-a5a8-4395-9793-2416c30721cd/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=63202) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 589.660551] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63202) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 589.660771] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.319s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 589.661052] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 5.456s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.699644] env[63202]: DEBUG nova.network.neutron [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 589.750259] env[63202]: DEBUG nova.virt.vmwareapi.images [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Downloaded image file data bb172aa4-a5a8-4395-9793-2416c30721cd to vmware_temp/b4c81966-06b3-4ef8-acd4-8bd374d92f76/bb172aa4-a5a8-4395-9793-2416c30721cd/tmp-sparse.vmdk on the data store datastore2 {{(pid=63202) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 589.750437] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Caching image {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 589.750735] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Copying Virtual Disk [datastore2] vmware_temp/b4c81966-06b3-4ef8-acd4-8bd374d92f76/bb172aa4-a5a8-4395-9793-2416c30721cd/tmp-sparse.vmdk to [datastore2] vmware_temp/b4c81966-06b3-4ef8-acd4-8bd374d92f76/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 589.752254] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-61426937-2a9a-4b21-a53d-a12b2eaecfe0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.760727] env[63202]: DEBUG oslo_vmware.api [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Waiting for the task: (returnval){ [ 589.760727] env[63202]: value = "task-1385292" [ 589.760727] env[63202]: _type = "Task" [ 589.760727] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 589.770392] env[63202]: DEBUG oslo_vmware.api [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385292, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 589.986340] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.010823] env[63202]: DEBUG nova.network.neutron [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.272086] env[63202]: DEBUG oslo_vmware.api [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385292, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 590.339722] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e9a4472-9155-4d7a-870b-9c498cee51de {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.351322] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd84683d-7d82-493f-adf9-d09917bdf4db {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.387952] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-088e8963-15b6-42b8-b9e3-69cde49e6cf1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.398703] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d368e67f-8de0-4f90-9af6-abf3c9c6276f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.414229] env[63202]: DEBUG nova.compute.provider_tree [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 590.518020] env[63202]: DEBUG oslo_concurrency.lockutils [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Releasing lock "refresh_cache-4646821e-13ec-4413-aa82-1bbf6eb384ad" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 590.518110] env[63202]: DEBUG nova.compute.manager [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 590.518303] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 590.518607] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9b7da8d2-6218-490a-8557-022a19ac1348 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.528780] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c4235aa-d75c-4c77-9317-621d68e223a1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.553709] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4646821e-13ec-4413-aa82-1bbf6eb384ad could not be found. [ 590.554016] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 590.554305] env[63202]: INFO nova.compute.manager [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] Took 0.04 seconds to destroy the instance on the hypervisor. [ 590.555238] env[63202]: DEBUG oslo.service.loopingcall [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 590.555436] env[63202]: DEBUG nova.compute.manager [-] [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 590.555606] env[63202]: DEBUG nova.network.neutron [-] [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 590.599921] env[63202]: DEBUG nova.network.neutron [-] [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 590.778212] env[63202]: DEBUG oslo_vmware.api [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385292, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.739388} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 590.778212] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Copied Virtual Disk [datastore2] vmware_temp/b4c81966-06b3-4ef8-acd4-8bd374d92f76/bb172aa4-a5a8-4395-9793-2416c30721cd/tmp-sparse.vmdk to [datastore2] vmware_temp/b4c81966-06b3-4ef8-acd4-8bd374d92f76/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 590.778212] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Deleting the datastore file [datastore2] vmware_temp/b4c81966-06b3-4ef8-acd4-8bd374d92f76/bb172aa4-a5a8-4395-9793-2416c30721cd/tmp-sparse.vmdk {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 590.778212] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5224145b-3fde-4aa0-8a05-dadf6f071ff2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.787625] env[63202]: DEBUG oslo_vmware.api [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Waiting for the task: (returnval){ [ 590.787625] env[63202]: value = "task-1385293" [ 590.787625] env[63202]: _type = "Task" [ 590.787625] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 590.799412] env[63202]: DEBUG oslo_vmware.api [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385293, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 590.831489] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] Acquiring lock "2f757f9b-537e-4bd8-b08d-1da991903552" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.831862] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] Lock "2f757f9b-537e-4bd8-b08d-1da991903552" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 590.917521] env[63202]: DEBUG nova.scheduler.client.report [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 591.101224] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Acquiring lock "c9b7182b-8116-419c-b399-2c7bcc133893" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.101224] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Lock "c9b7182b-8116-419c-b399-2c7bcc133893" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.105609] env[63202]: DEBUG nova.network.neutron [-] [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.155805] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Acquiring lock "4f0aba5d-1722-4ddb-8331-a10c25e29b6f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.155936] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Lock "4f0aba5d-1722-4ddb-8331-a10c25e29b6f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.164198] env[63202]: DEBUG nova.compute.manager [None req-50077aed-834b-4540-b4ae-d1fb15601799 tempest-ServerDiagnosticsV248Test-1611155863 tempest-ServerDiagnosticsV248Test-1611155863-project-admin] [instance: 21d0b283-3bb4-4bda-8e62-c933c6de4927] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 591.166382] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e84a5020-bf2c-4951-9888-83027dcb2f05 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.176970] env[63202]: INFO nova.compute.manager [None req-50077aed-834b-4540-b4ae-d1fb15601799 tempest-ServerDiagnosticsV248Test-1611155863 tempest-ServerDiagnosticsV248Test-1611155863-project-admin] [instance: 21d0b283-3bb4-4bda-8e62-c933c6de4927] Retrieving diagnostics [ 591.178540] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0213278-3e17-4ecc-928e-fb58dc5c1ae2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.301379] env[63202]: DEBUG oslo_vmware.api [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385293, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.031086} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 591.301839] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 591.302112] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Moving file from [datastore2] vmware_temp/b4c81966-06b3-4ef8-acd4-8bd374d92f76/bb172aa4-a5a8-4395-9793-2416c30721cd to [datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd. {{(pid=63202) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 591.302522] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-1664e7ec-b029-4d78-98a0-275cb43ae4c0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.311485] env[63202]: DEBUG oslo_vmware.api [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Waiting for the task: (returnval){ [ 591.311485] env[63202]: value = "task-1385294" [ 591.311485] env[63202]: _type = "Task" [ 591.311485] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 591.320609] env[63202]: DEBUG oslo_vmware.api [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385294, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 591.339323] env[63202]: DEBUG nova.compute.manager [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 591.428992] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.765s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 591.428992] env[63202]: ERROR nova.compute.manager [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] [instance: cc04e228-8b1a-4897-967e-131623434ac0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9f2cc00a-a922-474a-b98d-202cd3fb5d72, please check neutron logs for more information. [ 591.428992] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] Traceback (most recent call last): [ 591.428992] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 591.428992] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] self.driver.spawn(context, instance, image_meta, [ 591.428992] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 591.428992] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 591.428992] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 591.428992] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] vm_ref = self.build_virtual_machine(instance, [ 591.429305] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 591.429305] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] vif_infos = vmwarevif.get_vif_info(self._session, [ 591.429305] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 591.429305] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] for vif in network_info: [ 591.429305] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 591.429305] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] return self._sync_wrapper(fn, *args, **kwargs) [ 591.429305] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 591.429305] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] self.wait() [ 591.429305] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 591.429305] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] self[:] = self._gt.wait() [ 591.429305] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 591.429305] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] return self._exit_event.wait() [ 591.429305] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 591.429857] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] result = hub.switch() [ 591.429857] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 591.429857] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] return self.greenlet.switch() [ 591.429857] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 591.429857] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] result = function(*args, **kwargs) [ 591.429857] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 591.429857] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] return func(*args, **kwargs) [ 591.429857] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 591.429857] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] raise e [ 591.429857] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 591.429857] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] nwinfo = self.network_api.allocate_for_instance( [ 591.429857] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 591.429857] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] created_port_ids = self._update_ports_for_instance( [ 591.430230] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 591.430230] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] with excutils.save_and_reraise_exception(): [ 591.430230] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 591.430230] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] self.force_reraise() [ 591.430230] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 591.430230] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] raise self.value [ 591.430230] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 591.430230] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] updated_port = self._update_port( [ 591.430230] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 591.430230] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] _ensure_no_port_binding_failure(port) [ 591.430230] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 591.430230] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] raise exception.PortBindingFailed(port_id=port['id']) [ 591.430517] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] nova.exception.PortBindingFailed: Binding failed for port 9f2cc00a-a922-474a-b98d-202cd3fb5d72, please check neutron logs for more information. [ 591.430517] env[63202]: ERROR nova.compute.manager [instance: cc04e228-8b1a-4897-967e-131623434ac0] [ 591.430517] env[63202]: DEBUG nova.compute.utils [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] [instance: cc04e228-8b1a-4897-967e-131623434ac0] Binding failed for port 9f2cc00a-a922-474a-b98d-202cd3fb5d72, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 591.435120] env[63202]: DEBUG nova.compute.manager [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] [instance: cc04e228-8b1a-4897-967e-131623434ac0] Build of instance cc04e228-8b1a-4897-967e-131623434ac0 was re-scheduled: Binding failed for port 9f2cc00a-a922-474a-b98d-202cd3fb5d72, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 591.435120] env[63202]: DEBUG nova.compute.manager [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] [instance: cc04e228-8b1a-4897-967e-131623434ac0] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 591.435120] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] Acquiring lock "refresh_cache-cc04e228-8b1a-4897-967e-131623434ac0" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 591.435120] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] Acquired lock "refresh_cache-cc04e228-8b1a-4897-967e-131623434ac0" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 591.435391] env[63202]: DEBUG nova.network.neutron [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] [instance: cc04e228-8b1a-4897-967e-131623434ac0] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 591.437055] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 5.829s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.603447] env[63202]: DEBUG nova.compute.manager [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: c9b7182b-8116-419c-b399-2c7bcc133893] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 591.610648] env[63202]: INFO nova.compute.manager [-] [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] Took 1.05 seconds to deallocate network for instance. [ 591.613417] env[63202]: DEBUG nova.compute.claims [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 591.613417] env[63202]: DEBUG oslo_concurrency.lockutils [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.618597] env[63202]: ERROR nova.compute.manager [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5305bbd8-e698-4cb8-a36d-de3f0148ecac, please check neutron logs for more information. [ 591.618597] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 591.618597] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 591.618597] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 591.618597] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 591.618597] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 591.618597] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 591.618597] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 591.618597] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 591.618597] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 591.618597] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 591.618597] env[63202]: ERROR nova.compute.manager raise self.value [ 591.618597] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 591.618597] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 591.618597] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 591.618597] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 591.619178] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 591.619178] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 591.619178] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5305bbd8-e698-4cb8-a36d-de3f0148ecac, please check neutron logs for more information. [ 591.619178] env[63202]: ERROR nova.compute.manager [ 591.619178] env[63202]: Traceback (most recent call last): [ 591.619178] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 591.619178] env[63202]: listener.cb(fileno) [ 591.619178] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 591.619178] env[63202]: result = function(*args, **kwargs) [ 591.619178] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 591.619178] env[63202]: return func(*args, **kwargs) [ 591.619178] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 591.619178] env[63202]: raise e [ 591.619178] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 591.619178] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 591.619178] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 591.619178] env[63202]: created_port_ids = self._update_ports_for_instance( [ 591.619178] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 591.619178] env[63202]: with excutils.save_and_reraise_exception(): [ 591.619178] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 591.619178] env[63202]: self.force_reraise() [ 591.619178] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 591.619178] env[63202]: raise self.value [ 591.619178] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 591.619178] env[63202]: updated_port = self._update_port( [ 591.619178] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 591.619178] env[63202]: _ensure_no_port_binding_failure(port) [ 591.619178] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 591.619178] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 591.620014] env[63202]: nova.exception.PortBindingFailed: Binding failed for port 5305bbd8-e698-4cb8-a36d-de3f0148ecac, please check neutron logs for more information. [ 591.620014] env[63202]: Removing descriptor: 17 [ 591.620014] env[63202]: ERROR nova.compute.manager [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5305bbd8-e698-4cb8-a36d-de3f0148ecac, please check neutron logs for more information. [ 591.620014] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] Traceback (most recent call last): [ 591.620014] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 591.620014] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] yield resources [ 591.620014] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 591.620014] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] self.driver.spawn(context, instance, image_meta, [ 591.620014] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 591.620014] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 591.620014] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 591.620014] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] vm_ref = self.build_virtual_machine(instance, [ 591.621027] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 591.621027] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] vif_infos = vmwarevif.get_vif_info(self._session, [ 591.621027] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 591.621027] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] for vif in network_info: [ 591.621027] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 591.621027] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] return self._sync_wrapper(fn, *args, **kwargs) [ 591.621027] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 591.621027] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] self.wait() [ 591.621027] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 591.621027] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] self[:] = self._gt.wait() [ 591.621027] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 591.621027] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] return self._exit_event.wait() [ 591.621027] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 591.623904] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] result = hub.switch() [ 591.623904] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 591.623904] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] return self.greenlet.switch() [ 591.623904] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 591.623904] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] result = function(*args, **kwargs) [ 591.623904] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 591.623904] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] return func(*args, **kwargs) [ 591.623904] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 591.623904] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] raise e [ 591.623904] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 591.623904] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] nwinfo = self.network_api.allocate_for_instance( [ 591.623904] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 591.623904] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] created_port_ids = self._update_ports_for_instance( [ 591.624391] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 591.624391] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] with excutils.save_and_reraise_exception(): [ 591.624391] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 591.624391] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] self.force_reraise() [ 591.624391] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 591.624391] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] raise self.value [ 591.624391] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 591.624391] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] updated_port = self._update_port( [ 591.624391] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 591.624391] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] _ensure_no_port_binding_failure(port) [ 591.624391] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 591.624391] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] raise exception.PortBindingFailed(port_id=port['id']) [ 591.624754] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] nova.exception.PortBindingFailed: Binding failed for port 5305bbd8-e698-4cb8-a36d-de3f0148ecac, please check neutron logs for more information. [ 591.624754] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] [ 591.624754] env[63202]: INFO nova.compute.manager [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] Terminating instance [ 591.624754] env[63202]: DEBUG oslo_concurrency.lockutils [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] Acquiring lock "refresh_cache-6e102794-152e-492e-b063-e1ebaffa82b6" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 591.624754] env[63202]: DEBUG oslo_concurrency.lockutils [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] Acquired lock "refresh_cache-6e102794-152e-492e-b063-e1ebaffa82b6" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 591.624754] env[63202]: DEBUG nova.network.neutron [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 591.659425] env[63202]: DEBUG nova.compute.manager [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 591.823012] env[63202]: DEBUG oslo_vmware.api [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385294, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.023158} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 591.826438] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] File moved {{(pid=63202) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 591.826438] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Cleaning up location [datastore2] vmware_temp/b4c81966-06b3-4ef8-acd4-8bd374d92f76 {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 591.826438] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Deleting the datastore file [datastore2] vmware_temp/b4c81966-06b3-4ef8-acd4-8bd374d92f76 {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 591.826438] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-23a0f8b6-c414-4b61-824e-500dff48ad3d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.834194] env[63202]: DEBUG oslo_vmware.api [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Waiting for the task: (returnval){ [ 591.834194] env[63202]: value = "task-1385295" [ 591.834194] env[63202]: _type = "Task" [ 591.834194] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 591.848724] env[63202]: DEBUG oslo_vmware.api [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385295, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 591.875497] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.967878] env[63202]: DEBUG nova.network.neutron [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] [instance: cc04e228-8b1a-4897-967e-131623434ac0] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 592.125703] env[63202]: DEBUG nova.network.neutron [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] [instance: cc04e228-8b1a-4897-967e-131623434ac0] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 592.137341] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.151686] env[63202]: DEBUG nova.network.neutron [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 592.188334] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1e63fad-7180-43f8-9cdc-069de17e3c61 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.195438] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.200312] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c512a992-b45f-4817-9812-edbb11ce40d6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.236626] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38a07696-3cd8-4f06-8ef2-caa4395eee14 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.244529] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1729f91-a5af-45d4-afff-88ee6c52c40b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.259860] env[63202]: DEBUG nova.compute.provider_tree [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 592.310300] env[63202]: DEBUG nova.compute.manager [req-c1172a96-5e4b-4ebe-816f-52697a71feae req-34da448b-6065-416e-b618-03b176acc8e9 service nova] [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] Received event network-changed-f8672ac4-0bf6-410d-af1e-af7434600f7d {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 592.310300] env[63202]: DEBUG nova.compute.manager [req-c1172a96-5e4b-4ebe-816f-52697a71feae req-34da448b-6065-416e-b618-03b176acc8e9 service nova] [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] Refreshing instance network info cache due to event network-changed-f8672ac4-0bf6-410d-af1e-af7434600f7d. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 592.310300] env[63202]: DEBUG oslo_concurrency.lockutils [req-c1172a96-5e4b-4ebe-816f-52697a71feae req-34da448b-6065-416e-b618-03b176acc8e9 service nova] Acquiring lock "refresh_cache-4646821e-13ec-4413-aa82-1bbf6eb384ad" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 592.310420] env[63202]: DEBUG oslo_concurrency.lockutils [req-c1172a96-5e4b-4ebe-816f-52697a71feae req-34da448b-6065-416e-b618-03b176acc8e9 service nova] Acquired lock "refresh_cache-4646821e-13ec-4413-aa82-1bbf6eb384ad" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 592.310599] env[63202]: DEBUG nova.network.neutron [req-c1172a96-5e4b-4ebe-816f-52697a71feae req-34da448b-6065-416e-b618-03b176acc8e9 service nova] [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] Refreshing network info cache for port f8672ac4-0bf6-410d-af1e-af7434600f7d {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 592.343386] env[63202]: DEBUG nova.network.neutron [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 592.351517] env[63202]: DEBUG oslo_vmware.api [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385295, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.024981} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 592.351585] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 592.352341] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-605f4889-24c4-42a7-b506-6249522fc6fb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.358243] env[63202]: DEBUG oslo_vmware.api [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Waiting for the task: (returnval){ [ 592.358243] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52314270-f604-33c5-8a97-e53edbfafbd2" [ 592.358243] env[63202]: _type = "Task" [ 592.358243] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 592.369447] env[63202]: DEBUG oslo_vmware.api [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52314270-f604-33c5-8a97-e53edbfafbd2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 592.633615] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] Releasing lock "refresh_cache-cc04e228-8b1a-4897-967e-131623434ac0" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 592.634469] env[63202]: DEBUG nova.compute.manager [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 592.634731] env[63202]: DEBUG nova.compute.manager [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] [instance: cc04e228-8b1a-4897-967e-131623434ac0] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 592.635833] env[63202]: DEBUG nova.network.neutron [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] [instance: cc04e228-8b1a-4897-967e-131623434ac0] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 592.658152] env[63202]: DEBUG nova.network.neutron [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] [instance: cc04e228-8b1a-4897-967e-131623434ac0] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 592.764355] env[63202]: DEBUG nova.scheduler.client.report [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 592.836134] env[63202]: DEBUG oslo_concurrency.lockutils [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "29c289ea-703c-4083-b0c4-bbc8b4a92142" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.836391] env[63202]: DEBUG oslo_concurrency.lockutils [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "29c289ea-703c-4083-b0c4-bbc8b4a92142" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.847139] env[63202]: DEBUG oslo_concurrency.lockutils [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] Releasing lock "refresh_cache-6e102794-152e-492e-b063-e1ebaffa82b6" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 592.847139] env[63202]: DEBUG nova.compute.manager [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 592.847139] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 592.847598] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-167eea0f-8988-413d-a2a1-7fdeb61259e5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.865168] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fdf3aab-7f44-46e5-8351-82243063de31 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.875355] env[63202]: DEBUG nova.network.neutron [req-c1172a96-5e4b-4ebe-816f-52697a71feae req-34da448b-6065-416e-b618-03b176acc8e9 service nova] [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 592.888944] env[63202]: DEBUG oslo_vmware.api [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52314270-f604-33c5-8a97-e53edbfafbd2, 'name': SearchDatastore_Task, 'duration_secs': 0.008844} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 592.889219] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 592.889494] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore2] 0ded82aa-c68d-4d11-9e4d-90ed9a19b708/0ded82aa-c68d-4d11-9e4d-90ed9a19b708.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 592.889708] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-39109380-2ee4-4d2d-b9f7-37ef98720e29 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.900720] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6e102794-152e-492e-b063-e1ebaffa82b6 could not be found. [ 592.902192] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 592.902192] env[63202]: INFO nova.compute.manager [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] Took 0.05 seconds to destroy the instance on the hypervisor. [ 592.902192] env[63202]: DEBUG oslo.service.loopingcall [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 592.902192] env[63202]: DEBUG nova.compute.manager [-] [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 592.902192] env[63202]: DEBUG nova.network.neutron [-] [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 592.906636] env[63202]: DEBUG oslo_vmware.api [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Waiting for the task: (returnval){ [ 592.906636] env[63202]: value = "task-1385296" [ 592.906636] env[63202]: _type = "Task" [ 592.906636] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 592.917392] env[63202]: DEBUG oslo_vmware.api [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385296, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 592.974414] env[63202]: DEBUG nova.network.neutron [-] [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 593.069183] env[63202]: DEBUG nova.network.neutron [req-c1172a96-5e4b-4ebe-816f-52697a71feae req-34da448b-6065-416e-b618-03b176acc8e9 service nova] [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.161323] env[63202]: DEBUG nova.network.neutron [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] [instance: cc04e228-8b1a-4897-967e-131623434ac0] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.275352] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.839s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 593.275970] env[63202]: ERROR nova.compute.manager [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 77702c21-b059-4fe9-80ca-9fe5548801d9, please check neutron logs for more information. [ 593.275970] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] Traceback (most recent call last): [ 593.275970] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 593.275970] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] self.driver.spawn(context, instance, image_meta, [ 593.275970] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 593.275970] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 593.275970] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 593.275970] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] vm_ref = self.build_virtual_machine(instance, [ 593.275970] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 593.275970] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] vif_infos = vmwarevif.get_vif_info(self._session, [ 593.275970] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 593.276284] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] for vif in network_info: [ 593.276284] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 593.276284] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] return self._sync_wrapper(fn, *args, **kwargs) [ 593.276284] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 593.276284] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] self.wait() [ 593.276284] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 593.276284] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] self[:] = self._gt.wait() [ 593.276284] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 593.276284] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] return self._exit_event.wait() [ 593.276284] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 593.276284] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] result = hub.switch() [ 593.276284] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 593.276284] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] return self.greenlet.switch() [ 593.276613] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 593.276613] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] result = function(*args, **kwargs) [ 593.276613] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 593.276613] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] return func(*args, **kwargs) [ 593.276613] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 593.276613] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] raise e [ 593.276613] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 593.276613] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] nwinfo = self.network_api.allocate_for_instance( [ 593.276613] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 593.276613] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] created_port_ids = self._update_ports_for_instance( [ 593.276613] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 593.276613] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] with excutils.save_and_reraise_exception(): [ 593.276613] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 593.276957] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] self.force_reraise() [ 593.276957] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 593.276957] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] raise self.value [ 593.276957] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 593.276957] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] updated_port = self._update_port( [ 593.276957] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 593.276957] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] _ensure_no_port_binding_failure(port) [ 593.276957] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 593.276957] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] raise exception.PortBindingFailed(port_id=port['id']) [ 593.276957] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] nova.exception.PortBindingFailed: Binding failed for port 77702c21-b059-4fe9-80ca-9fe5548801d9, please check neutron logs for more information. [ 593.276957] env[63202]: ERROR nova.compute.manager [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] [ 593.277274] env[63202]: DEBUG nova.compute.utils [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] Binding failed for port 77702c21-b059-4fe9-80ca-9fe5548801d9, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 593.278442] env[63202]: DEBUG oslo_concurrency.lockutils [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.909s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.280116] env[63202]: INFO nova.compute.claims [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 593.283770] env[63202]: DEBUG nova.compute.manager [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] Build of instance 4a0241c3-8dea-4deb-80e1-66bb9cba10b9 was re-scheduled: Binding failed for port 77702c21-b059-4fe9-80ca-9fe5548801d9, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 593.284446] env[63202]: DEBUG nova.compute.manager [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 593.284559] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] Acquiring lock "refresh_cache-4a0241c3-8dea-4deb-80e1-66bb9cba10b9" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 593.284707] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] Acquired lock "refresh_cache-4a0241c3-8dea-4deb-80e1-66bb9cba10b9" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 593.284866] env[63202]: DEBUG nova.network.neutron [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 593.339114] env[63202]: DEBUG nova.compute.manager [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 593.425262] env[63202]: DEBUG oslo_vmware.api [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385296, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.494639} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 593.425744] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore2] 0ded82aa-c68d-4d11-9e4d-90ed9a19b708/0ded82aa-c68d-4d11-9e4d-90ed9a19b708.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 593.426261] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 593.428069] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e1f6c406-3e7e-4a91-a6b1-108e89f7c277 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.435838] env[63202]: DEBUG oslo_vmware.api [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Waiting for the task: (returnval){ [ 593.435838] env[63202]: value = "task-1385297" [ 593.435838] env[63202]: _type = "Task" [ 593.435838] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 593.449333] env[63202]: DEBUG oslo_vmware.api [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385297, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 593.484788] env[63202]: DEBUG nova.network.neutron [-] [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.576267] env[63202]: DEBUG oslo_concurrency.lockutils [req-c1172a96-5e4b-4ebe-816f-52697a71feae req-34da448b-6065-416e-b618-03b176acc8e9 service nova] Releasing lock "refresh_cache-4646821e-13ec-4413-aa82-1bbf6eb384ad" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 593.577022] env[63202]: DEBUG nova.compute.manager [req-c1172a96-5e4b-4ebe-816f-52697a71feae req-34da448b-6065-416e-b618-03b176acc8e9 service nova] [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] Received event network-vif-deleted-f8672ac4-0bf6-410d-af1e-af7434600f7d {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 593.665953] env[63202]: INFO nova.compute.manager [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] [instance: cc04e228-8b1a-4897-967e-131623434ac0] Took 1.03 seconds to deallocate network for instance. [ 593.824346] env[63202]: DEBUG nova.network.neutron [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 593.872755] env[63202]: DEBUG oslo_concurrency.lockutils [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 593.950466] env[63202]: DEBUG oslo_vmware.api [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385297, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06742} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 593.950466] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 593.951984] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-738e69f1-e467-4235-a7cd-3f82424e8e36 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.976154] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Reconfiguring VM instance instance-00000002 to attach disk [datastore2] 0ded82aa-c68d-4d11-9e4d-90ed9a19b708/0ded82aa-c68d-4d11-9e4d-90ed9a19b708.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 593.976389] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-608cd690-b0d4-4c9d-8923-c712086319b2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.996985] env[63202]: INFO nova.compute.manager [-] [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] Took 1.09 seconds to deallocate network for instance. [ 593.998213] env[63202]: DEBUG nova.compute.claims [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 593.998384] env[63202]: DEBUG oslo_concurrency.lockutils [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.004179] env[63202]: DEBUG oslo_vmware.api [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Waiting for the task: (returnval){ [ 594.004179] env[63202]: value = "task-1385298" [ 594.004179] env[63202]: _type = "Task" [ 594.004179] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 594.014490] env[63202]: DEBUG oslo_vmware.api [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385298, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 594.206430] env[63202]: DEBUG nova.network.neutron [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.515098] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fab0be72-5bf4-4183-a175-c2465ee7f0ca {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.520917] env[63202]: DEBUG oslo_vmware.api [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385298, 'name': ReconfigVM_Task, 'duration_secs': 0.280053} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 594.521602] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Reconfigured VM instance instance-00000002 to attach disk [datastore2] 0ded82aa-c68d-4d11-9e4d-90ed9a19b708/0ded82aa-c68d-4d11-9e4d-90ed9a19b708.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 594.522232] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e45581d9-24af-4c1f-b7db-00b542d072e9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.526539] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f667f3d-a24f-4bbd-ad6c-1abae6754f32 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.530667] env[63202]: DEBUG oslo_vmware.api [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Waiting for the task: (returnval){ [ 594.530667] env[63202]: value = "task-1385299" [ 594.530667] env[63202]: _type = "Task" [ 594.530667] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 594.558525] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a285724-87e4-4e6e-aefe-53c53f7a04af {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.564404] env[63202]: DEBUG oslo_vmware.api [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385299, 'name': Rename_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 594.569317] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4833398a-dc43-4f8a-885b-7edebf1cad39 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.582760] env[63202]: DEBUG nova.compute.provider_tree [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 594.715173] env[63202]: INFO nova.scheduler.client.report [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] Deleted allocations for instance cc04e228-8b1a-4897-967e-131623434ac0 [ 594.721302] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] Releasing lock "refresh_cache-4a0241c3-8dea-4deb-80e1-66bb9cba10b9" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 594.722395] env[63202]: DEBUG nova.compute.manager [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 594.722395] env[63202]: DEBUG nova.compute.manager [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 594.722395] env[63202]: DEBUG nova.network.neutron [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 594.759219] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Acquiring lock "88f60509-35cc-4d70-a97c-76dd2761c89b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.759834] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Lock "88f60509-35cc-4d70-a97c-76dd2761c89b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.780787] env[63202]: DEBUG nova.network.neutron [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 595.044764] env[63202]: DEBUG oslo_vmware.api [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385299, 'name': Rename_Task, 'duration_secs': 0.388637} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 595.045381] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 595.045726] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cf81c25d-a8d9-42ec-baa2-a63e449edbef {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.055060] env[63202]: DEBUG oslo_vmware.api [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Waiting for the task: (returnval){ [ 595.055060] env[63202]: value = "task-1385300" [ 595.055060] env[63202]: _type = "Task" [ 595.055060] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 595.065534] env[63202]: DEBUG oslo_vmware.api [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385300, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 595.086573] env[63202]: DEBUG nova.scheduler.client.report [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 595.226895] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d6d98d31-c617-4b01-9233-663c2836bcee tempest-ImagesOneServerTestJSON-1514129301 tempest-ImagesOneServerTestJSON-1514129301-project-member] Lock "cc04e228-8b1a-4897-967e-131623434ac0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.145s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 595.284234] env[63202]: DEBUG nova.network.neutron [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 595.570985] env[63202]: DEBUG oslo_vmware.api [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385300, 'name': PowerOnVM_Task} progress is 80%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 595.595461] env[63202]: DEBUG oslo_concurrency.lockutils [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.316s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 595.595567] env[63202]: DEBUG nova.compute.manager [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 595.601332] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.612s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.603305] env[63202]: INFO nova.compute.claims [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 595.733038] env[63202]: DEBUG nova.compute.manager [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 595.791092] env[63202]: INFO nova.compute.manager [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] [instance: 4a0241c3-8dea-4deb-80e1-66bb9cba10b9] Took 1.07 seconds to deallocate network for instance. [ 596.075743] env[63202]: DEBUG oslo_vmware.api [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385300, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 596.108368] env[63202]: DEBUG nova.compute.utils [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 596.113544] env[63202]: DEBUG nova.compute.manager [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 596.113785] env[63202]: DEBUG nova.network.neutron [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 596.213438] env[63202]: DEBUG nova.policy [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '90f57a662ade4cceab48c73ab07e7ca1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1f5e5ca683014ba7929b26cdd2bce441', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 596.225552] env[63202]: DEBUG nova.compute.manager [req-8f865e72-56e6-4c62-8719-8c090069b1b4 req-52594589-cd24-41e2-a66a-d3d699f2b0ec service nova] [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] Received event network-changed-5305bbd8-e698-4cb8-a36d-de3f0148ecac {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 596.225621] env[63202]: DEBUG nova.compute.manager [req-8f865e72-56e6-4c62-8719-8c090069b1b4 req-52594589-cd24-41e2-a66a-d3d699f2b0ec service nova] [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] Refreshing instance network info cache due to event network-changed-5305bbd8-e698-4cb8-a36d-de3f0148ecac. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 596.225858] env[63202]: DEBUG oslo_concurrency.lockutils [req-8f865e72-56e6-4c62-8719-8c090069b1b4 req-52594589-cd24-41e2-a66a-d3d699f2b0ec service nova] Acquiring lock "refresh_cache-6e102794-152e-492e-b063-e1ebaffa82b6" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 596.225982] env[63202]: DEBUG oslo_concurrency.lockutils [req-8f865e72-56e6-4c62-8719-8c090069b1b4 req-52594589-cd24-41e2-a66a-d3d699f2b0ec service nova] Acquired lock "refresh_cache-6e102794-152e-492e-b063-e1ebaffa82b6" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 596.226167] env[63202]: DEBUG nova.network.neutron [req-8f865e72-56e6-4c62-8719-8c090069b1b4 req-52594589-cd24-41e2-a66a-d3d699f2b0ec service nova] [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] Refreshing network info cache for port 5305bbd8-e698-4cb8-a36d-de3f0148ecac {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 596.281815] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.567844] env[63202]: DEBUG oslo_vmware.api [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385300, 'name': PowerOnVM_Task, 'duration_secs': 1.072532} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 596.568354] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 596.568671] env[63202]: DEBUG nova.compute.manager [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 596.569794] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c959261-67bc-4c36-8615-d534607f5c14 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.621815] env[63202]: DEBUG nova.compute.manager [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 596.775554] env[63202]: DEBUG nova.network.neutron [req-8f865e72-56e6-4c62-8719-8c090069b1b4 req-52594589-cd24-41e2-a66a-d3d699f2b0ec service nova] [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 596.842245] env[63202]: INFO nova.scheduler.client.report [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] Deleted allocations for instance 4a0241c3-8dea-4deb-80e1-66bb9cba10b9 [ 596.936368] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] Acquiring lock "1c4fe9af-7a19-4aaf-8758-8d945b2ee57e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.936615] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] Lock "1c4fe9af-7a19-4aaf-8758-8d945b2ee57e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 596.937538] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdf8a317-eea2-412a-ad6f-8860b2854e38 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.948577] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4538c529-8d98-4f17-aba2-a4da8fedbcda {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.985962] env[63202]: DEBUG nova.network.neutron [req-8f865e72-56e6-4c62-8719-8c090069b1b4 req-52594589-cd24-41e2-a66a-d3d699f2b0ec service nova] [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.987684] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-521ae144-e480-43c7-86d6-e8368baee279 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.995776] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2a3ca66-cd97-43dc-887b-342f6facd895 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.009999] env[63202]: DEBUG nova.compute.provider_tree [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 597.092532] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.238681] env[63202]: DEBUG nova.network.neutron [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] Successfully created port: 180d9c18-47ea-4305-a5b8-ab5abcdf5a1a {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 597.355407] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d208a85e-3293-4b12-8cda-70b437db5255 tempest-FloatingIPsAssociationTestJSON-62671885 tempest-FloatingIPsAssociationTestJSON-62671885-project-member] Lock "4a0241c3-8dea-4deb-80e1-66bb9cba10b9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.270s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 597.493016] env[63202]: DEBUG oslo_concurrency.lockutils [req-8f865e72-56e6-4c62-8719-8c090069b1b4 req-52594589-cd24-41e2-a66a-d3d699f2b0ec service nova] Releasing lock "refresh_cache-6e102794-152e-492e-b063-e1ebaffa82b6" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 597.493016] env[63202]: DEBUG nova.compute.manager [req-8f865e72-56e6-4c62-8719-8c090069b1b4 req-52594589-cd24-41e2-a66a-d3d699f2b0ec service nova] [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] Received event network-vif-deleted-5305bbd8-e698-4cb8-a36d-de3f0148ecac {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 597.516934] env[63202]: DEBUG nova.scheduler.client.report [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 597.641933] env[63202]: DEBUG nova.compute.manager [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 597.670633] env[63202]: DEBUG nova.virt.hardware [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 597.670868] env[63202]: DEBUG nova.virt.hardware [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 597.671030] env[63202]: DEBUG nova.virt.hardware [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 597.671215] env[63202]: DEBUG nova.virt.hardware [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 597.671357] env[63202]: DEBUG nova.virt.hardware [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 597.671497] env[63202]: DEBUG nova.virt.hardware [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 597.671697] env[63202]: DEBUG nova.virt.hardware [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 597.671849] env[63202]: DEBUG nova.virt.hardware [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 597.672015] env[63202]: DEBUG nova.virt.hardware [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 597.672297] env[63202]: DEBUG nova.virt.hardware [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 597.672511] env[63202]: DEBUG nova.virt.hardware [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 597.673405] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8d6d0ef-0413-4bec-a692-b8f9a818033a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.682300] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4517476-2b44-473c-b432-e1ba9133d2da {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.861012] env[63202]: DEBUG nova.compute.manager [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 598.022381] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.424s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 598.022937] env[63202]: DEBUG nova.compute.manager [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 598.027273] env[63202]: DEBUG oslo_concurrency.lockutils [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 6.414s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.386381] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.533428] env[63202]: DEBUG nova.compute.utils [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 598.538611] env[63202]: DEBUG nova.compute.manager [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 598.538797] env[63202]: DEBUG nova.network.neutron [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 598.640783] env[63202]: DEBUG nova.policy [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e8f61e346d654c4399b59c59e94ed227', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '777b9af00b5145a99c726dcc9b5f8c22', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 598.731725] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c1d609a-ac3e-4f1b-89f3-7b21732bf334 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.741969] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b13cdbb-ffc7-48d7-9320-1875c9b19aba {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.778091] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-effbc4fd-a279-417a-946f-5c8aea1c03ed {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.785790] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fa124a6-27c1-4cbe-9f6d-c6bb289f2190 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.802156] env[63202]: DEBUG nova.compute.provider_tree [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 599.039402] env[63202]: DEBUG nova.compute.manager [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 599.278174] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] Acquiring lock "1afb7b0e-7c9a-4594-8214-37fbecc1b7c5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 599.278174] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] Lock "1afb7b0e-7c9a-4594-8214-37fbecc1b7c5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 599.304586] env[63202]: DEBUG nova.scheduler.client.report [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 599.754868] env[63202]: DEBUG nova.network.neutron [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] Successfully created port: 684d1759-1f66-45f5-86c1-6fad8a310af1 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 599.814518] env[63202]: DEBUG oslo_concurrency.lockutils [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.784s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 599.814518] env[63202]: ERROR nova.compute.manager [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f8672ac4-0bf6-410d-af1e-af7434600f7d, please check neutron logs for more information. [ 599.814518] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] Traceback (most recent call last): [ 599.814518] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 599.814518] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] self.driver.spawn(context, instance, image_meta, [ 599.814518] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 599.814518] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] self._vmops.spawn(context, instance, image_meta, injected_files, [ 599.814518] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 599.814518] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] vm_ref = self.build_virtual_machine(instance, [ 599.815076] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 599.815076] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] vif_infos = vmwarevif.get_vif_info(self._session, [ 599.815076] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 599.815076] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] for vif in network_info: [ 599.815076] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 599.815076] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] return self._sync_wrapper(fn, *args, **kwargs) [ 599.815076] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 599.815076] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] self.wait() [ 599.815076] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 599.815076] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] self[:] = self._gt.wait() [ 599.815076] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 599.815076] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] return self._exit_event.wait() [ 599.815076] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 599.815391] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] result = hub.switch() [ 599.815391] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 599.815391] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] return self.greenlet.switch() [ 599.815391] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 599.815391] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] result = function(*args, **kwargs) [ 599.815391] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 599.815391] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] return func(*args, **kwargs) [ 599.815391] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 599.815391] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] raise e [ 599.815391] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 599.815391] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] nwinfo = self.network_api.allocate_for_instance( [ 599.815391] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 599.815391] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] created_port_ids = self._update_ports_for_instance( [ 599.815694] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 599.815694] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] with excutils.save_and_reraise_exception(): [ 599.815694] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 599.815694] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] self.force_reraise() [ 599.815694] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 599.815694] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] raise self.value [ 599.815694] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 599.815694] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] updated_port = self._update_port( [ 599.815694] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 599.815694] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] _ensure_no_port_binding_failure(port) [ 599.815694] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 599.815694] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] raise exception.PortBindingFailed(port_id=port['id']) [ 599.815985] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] nova.exception.PortBindingFailed: Binding failed for port f8672ac4-0bf6-410d-af1e-af7434600f7d, please check neutron logs for more information. [ 599.815985] env[63202]: ERROR nova.compute.manager [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] [ 599.815985] env[63202]: DEBUG nova.compute.utils [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] Binding failed for port f8672ac4-0bf6-410d-af1e-af7434600f7d, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 599.815985] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.938s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 599.817709] env[63202]: INFO nova.compute.claims [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 599.820613] env[63202]: DEBUG nova.compute.manager [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] Build of instance 4646821e-13ec-4413-aa82-1bbf6eb384ad was re-scheduled: Binding failed for port f8672ac4-0bf6-410d-af1e-af7434600f7d, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 599.821068] env[63202]: DEBUG nova.compute.manager [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 599.821436] env[63202]: DEBUG oslo_concurrency.lockutils [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Acquiring lock "refresh_cache-4646821e-13ec-4413-aa82-1bbf6eb384ad" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 599.821608] env[63202]: DEBUG oslo_concurrency.lockutils [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Acquired lock "refresh_cache-4646821e-13ec-4413-aa82-1bbf6eb384ad" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 599.821938] env[63202]: DEBUG nova.network.neutron [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 600.049013] env[63202]: DEBUG nova.compute.manager [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 600.085038] env[63202]: DEBUG nova.virt.hardware [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 600.085038] env[63202]: DEBUG nova.virt.hardware [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 600.085038] env[63202]: DEBUG nova.virt.hardware [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 600.085038] env[63202]: DEBUG nova.virt.hardware [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 600.085287] env[63202]: DEBUG nova.virt.hardware [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 600.085287] env[63202]: DEBUG nova.virt.hardware [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 600.085362] env[63202]: DEBUG nova.virt.hardware [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 600.085516] env[63202]: DEBUG nova.virt.hardware [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 600.086077] env[63202]: DEBUG nova.virt.hardware [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 600.086077] env[63202]: DEBUG nova.virt.hardware [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 600.086077] env[63202]: DEBUG nova.virt.hardware [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 600.087074] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f2a03be-5bfb-4f23-9014-d21f5e90da43 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.096585] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e33b87ea-f2e5-49d0-91dd-3d468adf1457 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.357800] env[63202]: DEBUG nova.network.neutron [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 600.555531] env[63202]: DEBUG nova.network.neutron [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.966687] env[63202]: INFO nova.compute.manager [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Rebuilding instance [ 600.969852] env[63202]: DEBUG nova.network.neutron [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] Successfully created port: 97e0d874-2c9e-476a-aa4f-2a0b38d61c4f {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 601.020192] env[63202]: DEBUG nova.compute.manager [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 601.022659] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b7b3c33-5563-4dbb-ac36-d6024e7eb31e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.059702] env[63202]: DEBUG oslo_concurrency.lockutils [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Releasing lock "refresh_cache-4646821e-13ec-4413-aa82-1bbf6eb384ad" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 601.059964] env[63202]: DEBUG nova.compute.manager [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 601.060206] env[63202]: DEBUG nova.compute.manager [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 601.060378] env[63202]: DEBUG nova.network.neutron [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 601.092157] env[63202]: DEBUG nova.network.neutron [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 601.112303] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e0af342-3a64-4546-8ca8-58d8d2ce37d6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.122308] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b01d4d61-1e9c-4e0a-8503-58393c5c5467 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.157579] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f7c388d-c37f-4340-a592-e5cf66a13f28 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.166453] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57912d2d-d96d-4d23-a7c0-ea3a520b0e94 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.182429] env[63202]: DEBUG nova.compute.provider_tree [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 601.539394] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 601.539974] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-485ca952-94ca-4346-a055-e62fdc5eb02e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.547335] env[63202]: DEBUG oslo_vmware.api [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Waiting for the task: (returnval){ [ 601.547335] env[63202]: value = "task-1385301" [ 601.547335] env[63202]: _type = "Task" [ 601.547335] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 601.563172] env[63202]: DEBUG oslo_vmware.api [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Task: {'id': task-1385301, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 601.598643] env[63202]: DEBUG nova.network.neutron [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.606276] env[63202]: ERROR nova.compute.manager [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 180d9c18-47ea-4305-a5b8-ab5abcdf5a1a, please check neutron logs for more information. [ 601.606276] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 601.606276] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 601.606276] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 601.606276] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 601.606276] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 601.606276] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 601.606276] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 601.606276] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.606276] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 601.606276] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.606276] env[63202]: ERROR nova.compute.manager raise self.value [ 601.606276] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 601.606276] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 601.606276] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.606276] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 601.606797] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.606797] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 601.606797] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 180d9c18-47ea-4305-a5b8-ab5abcdf5a1a, please check neutron logs for more information. [ 601.606797] env[63202]: ERROR nova.compute.manager [ 601.606797] env[63202]: Traceback (most recent call last): [ 601.606797] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 601.606797] env[63202]: listener.cb(fileno) [ 601.606797] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 601.606797] env[63202]: result = function(*args, **kwargs) [ 601.606797] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 601.606797] env[63202]: return func(*args, **kwargs) [ 601.606797] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 601.606797] env[63202]: raise e [ 601.606797] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 601.606797] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 601.606797] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 601.606797] env[63202]: created_port_ids = self._update_ports_for_instance( [ 601.606797] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 601.606797] env[63202]: with excutils.save_and_reraise_exception(): [ 601.606797] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.606797] env[63202]: self.force_reraise() [ 601.606797] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.606797] env[63202]: raise self.value [ 601.606797] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 601.606797] env[63202]: updated_port = self._update_port( [ 601.606797] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.606797] env[63202]: _ensure_no_port_binding_failure(port) [ 601.606797] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.606797] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 601.607688] env[63202]: nova.exception.PortBindingFailed: Binding failed for port 180d9c18-47ea-4305-a5b8-ab5abcdf5a1a, please check neutron logs for more information. [ 601.607688] env[63202]: Removing descriptor: 17 [ 601.607688] env[63202]: ERROR nova.compute.manager [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 180d9c18-47ea-4305-a5b8-ab5abcdf5a1a, please check neutron logs for more information. [ 601.607688] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] Traceback (most recent call last): [ 601.607688] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 601.607688] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] yield resources [ 601.607688] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 601.607688] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] self.driver.spawn(context, instance, image_meta, [ 601.607688] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 601.607688] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 601.607688] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 601.607688] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] vm_ref = self.build_virtual_machine(instance, [ 601.607994] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 601.607994] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] vif_infos = vmwarevif.get_vif_info(self._session, [ 601.607994] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 601.607994] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] for vif in network_info: [ 601.607994] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 601.607994] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] return self._sync_wrapper(fn, *args, **kwargs) [ 601.607994] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 601.607994] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] self.wait() [ 601.607994] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 601.607994] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] self[:] = self._gt.wait() [ 601.607994] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 601.607994] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] return self._exit_event.wait() [ 601.607994] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 601.608368] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] result = hub.switch() [ 601.608368] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 601.608368] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] return self.greenlet.switch() [ 601.608368] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 601.608368] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] result = function(*args, **kwargs) [ 601.608368] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 601.608368] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] return func(*args, **kwargs) [ 601.608368] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 601.608368] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] raise e [ 601.608368] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 601.608368] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] nwinfo = self.network_api.allocate_for_instance( [ 601.608368] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 601.608368] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] created_port_ids = self._update_ports_for_instance( [ 601.608716] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 601.608716] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] with excutils.save_and_reraise_exception(): [ 601.608716] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.608716] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] self.force_reraise() [ 601.608716] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.608716] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] raise self.value [ 601.608716] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 601.608716] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] updated_port = self._update_port( [ 601.608716] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.608716] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] _ensure_no_port_binding_failure(port) [ 601.608716] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.608716] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] raise exception.PortBindingFailed(port_id=port['id']) [ 601.609235] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] nova.exception.PortBindingFailed: Binding failed for port 180d9c18-47ea-4305-a5b8-ab5abcdf5a1a, please check neutron logs for more information. [ 601.609235] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] [ 601.609235] env[63202]: INFO nova.compute.manager [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] Terminating instance [ 601.609235] env[63202]: DEBUG oslo_concurrency.lockutils [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] Acquiring lock "refresh_cache-90999eb1-8829-4f99-b3b9-8c70ee636cc5" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 601.609235] env[63202]: DEBUG oslo_concurrency.lockutils [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] Acquired lock "refresh_cache-90999eb1-8829-4f99-b3b9-8c70ee636cc5" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.609235] env[63202]: DEBUG nova.network.neutron [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 601.685581] env[63202]: DEBUG nova.scheduler.client.report [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 601.815802] env[63202]: DEBUG nova.network.neutron [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] Successfully created port: c5f443af-9713-469d-af72-895e7c704075 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 602.061696] env[63202]: DEBUG oslo_vmware.api [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Task: {'id': task-1385301, 'name': PowerOffVM_Task, 'duration_secs': 0.14531} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 602.062409] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 602.062590] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 602.063935] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-963db97b-bff7-4ec6-b9a1-206e384d261e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.073123] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 602.073373] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a77ef084-e831-4233-a60a-a8d2767afe6a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.101844] env[63202]: INFO nova.compute.manager [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 4646821e-13ec-4413-aa82-1bbf6eb384ad] Took 1.04 seconds to deallocate network for instance. [ 602.106158] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 602.108179] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Deleting contents of the VM from datastore datastore2 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 602.108179] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Deleting the datastore file [datastore2] 0ded82aa-c68d-4d11-9e4d-90ed9a19b708 {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 602.108179] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ddaec34d-c13e-493f-81bb-0825afaa5c74 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.120187] env[63202]: DEBUG oslo_vmware.api [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Waiting for the task: (returnval){ [ 602.120187] env[63202]: value = "task-1385303" [ 602.120187] env[63202]: _type = "Task" [ 602.120187] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 602.133052] env[63202]: DEBUG oslo_vmware.api [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Task: {'id': task-1385303, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 602.183299] env[63202]: DEBUG nova.network.neutron [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 602.191134] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.377s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.191645] env[63202]: DEBUG nova.compute.manager [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 602.194461] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.057s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.195968] env[63202]: INFO nova.compute.claims [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: c9b7182b-8116-419c-b399-2c7bcc133893] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 602.493865] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquiring lock "13e91ad9-7fe6-4bbe-8a67-d63a48bfac20" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.494106] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lock "13e91ad9-7fe6-4bbe-8a67-d63a48bfac20" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.632010] env[63202]: DEBUG oslo_vmware.api [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Task: {'id': task-1385303, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.10874} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 602.632276] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 602.632459] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Deleted contents of the VM from datastore datastore2 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 602.632630] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 602.700382] env[63202]: DEBUG nova.compute.utils [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 602.703643] env[63202]: DEBUG nova.network.neutron [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.708887] env[63202]: DEBUG nova.compute.manager [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 602.709543] env[63202]: DEBUG nova.network.neutron [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 602.878727] env[63202]: DEBUG nova.policy [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1f41a89c5d2e44608554e47608ba86eb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b1702688161342908eb5fbe928d47e52', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 603.157127] env[63202]: DEBUG nova.compute.manager [req-0bd2a14b-9fc5-41fb-9802-294ab7216db9 req-f439376e-fefe-4bc9-bc1c-888e30005a8c service nova] [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] Received event network-changed-180d9c18-47ea-4305-a5b8-ab5abcdf5a1a {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 603.158289] env[63202]: DEBUG nova.compute.manager [req-0bd2a14b-9fc5-41fb-9802-294ab7216db9 req-f439376e-fefe-4bc9-bc1c-888e30005a8c service nova] [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] Refreshing instance network info cache due to event network-changed-180d9c18-47ea-4305-a5b8-ab5abcdf5a1a. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 603.159864] env[63202]: DEBUG oslo_concurrency.lockutils [req-0bd2a14b-9fc5-41fb-9802-294ab7216db9 req-f439376e-fefe-4bc9-bc1c-888e30005a8c service nova] Acquiring lock "refresh_cache-90999eb1-8829-4f99-b3b9-8c70ee636cc5" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 603.162215] env[63202]: INFO nova.scheduler.client.report [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Deleted allocations for instance 4646821e-13ec-4413-aa82-1bbf6eb384ad [ 603.215330] env[63202]: DEBUG nova.compute.manager [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 603.216571] env[63202]: DEBUG oslo_concurrency.lockutils [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] Releasing lock "refresh_cache-90999eb1-8829-4f99-b3b9-8c70ee636cc5" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.217132] env[63202]: DEBUG nova.compute.manager [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 603.217467] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 603.221358] env[63202]: DEBUG oslo_concurrency.lockutils [req-0bd2a14b-9fc5-41fb-9802-294ab7216db9 req-f439376e-fefe-4bc9-bc1c-888e30005a8c service nova] Acquired lock "refresh_cache-90999eb1-8829-4f99-b3b9-8c70ee636cc5" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 603.221686] env[63202]: DEBUG nova.network.neutron [req-0bd2a14b-9fc5-41fb-9802-294ab7216db9 req-f439376e-fefe-4bc9-bc1c-888e30005a8c service nova] [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] Refreshing network info cache for port 180d9c18-47ea-4305-a5b8-ab5abcdf5a1a {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 603.223537] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-267c4acf-77c2-48b7-a4d5-49fbcef6233d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.234581] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58a5bc7b-24a1-435a-a67f-d9fc4b18b94e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.262131] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 90999eb1-8829-4f99-b3b9-8c70ee636cc5 could not be found. [ 603.262131] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 603.262131] env[63202]: INFO nova.compute.manager [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] Took 0.04 seconds to destroy the instance on the hypervisor. [ 603.262570] env[63202]: DEBUG oslo.service.loopingcall [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 603.262570] env[63202]: DEBUG nova.compute.manager [-] [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 603.263373] env[63202]: DEBUG nova.network.neutron [-] [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 603.291694] env[63202]: DEBUG nova.network.neutron [-] [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 603.456546] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a4871cb-f5f9-4d81-aba3-e1ae27c408d8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.464874] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e53a571e-7dec-425e-af80-0206d1ae60eb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.499290] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fc58fe0-7f14-4294-bfa1-9b8eb873836b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.507091] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61b33c41-00f4-4ec5-83fe-69bb027b07d8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.520794] env[63202]: DEBUG nova.compute.provider_tree [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 603.672940] env[63202]: DEBUG oslo_concurrency.lockutils [None req-87197476-4580-47f0-a72b-6dd036e48557 tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Lock "4646821e-13ec-4413-aa82-1bbf6eb384ad" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.178s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 603.680113] env[63202]: DEBUG nova.virt.hardware [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 603.680285] env[63202]: DEBUG nova.virt.hardware [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 603.680339] env[63202]: DEBUG nova.virt.hardware [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 603.680741] env[63202]: DEBUG nova.virt.hardware [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 603.681504] env[63202]: DEBUG nova.virt.hardware [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 603.681504] env[63202]: DEBUG nova.virt.hardware [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 603.681588] env[63202]: DEBUG nova.virt.hardware [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 603.682043] env[63202]: DEBUG nova.virt.hardware [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 603.682389] env[63202]: DEBUG nova.virt.hardware [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 603.683740] env[63202]: DEBUG nova.virt.hardware [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 603.683740] env[63202]: DEBUG nova.virt.hardware [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 603.683978] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff63319e-a163-4eb2-9f5a-84f1fc1425e0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.697208] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d535d5a5-6db3-4c7f-ab9f-018ba025e3ab {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.710470] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Instance VIF info [] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 603.719078] env[63202]: DEBUG oslo.service.loopingcall [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 603.719078] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 603.719078] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8dce4fc4-b376-4cf4-a4e7-1d229b382749 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.743851] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 603.743851] env[63202]: value = "task-1385304" [ 603.743851] env[63202]: _type = "Task" [ 603.743851] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 603.755123] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385304, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 603.762047] env[63202]: DEBUG nova.network.neutron [req-0bd2a14b-9fc5-41fb-9802-294ab7216db9 req-f439376e-fefe-4bc9-bc1c-888e30005a8c service nova] [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 603.794575] env[63202]: DEBUG nova.network.neutron [-] [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.895297] env[63202]: DEBUG nova.network.neutron [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] Successfully created port: b35498a4-9923-43db-8dda-1e3802f5777a {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 603.954176] env[63202]: DEBUG nova.network.neutron [req-0bd2a14b-9fc5-41fb-9802-294ab7216db9 req-f439376e-fefe-4bc9-bc1c-888e30005a8c service nova] [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.025157] env[63202]: DEBUG nova.scheduler.client.report [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 604.176035] env[63202]: DEBUG nova.compute.manager [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 604.235352] env[63202]: DEBUG nova.compute.manager [None req-e99a975b-cde0-4dcd-868c-5fcfd93aea73 tempest-ServerDiagnosticsV248Test-1611155863 tempest-ServerDiagnosticsV248Test-1611155863-project-admin] [instance: 21d0b283-3bb4-4bda-8e62-c933c6de4927] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 604.236856] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8c1326d-668c-4faa-8b53-e1d16bca803d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.240908] env[63202]: DEBUG nova.compute.manager [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 604.250519] env[63202]: INFO nova.compute.manager [None req-e99a975b-cde0-4dcd-868c-5fcfd93aea73 tempest-ServerDiagnosticsV248Test-1611155863 tempest-ServerDiagnosticsV248Test-1611155863-project-admin] [instance: 21d0b283-3bb4-4bda-8e62-c933c6de4927] Retrieving diagnostics [ 604.252103] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a3eed02-98cd-47df-8a2b-7097324bb370 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.259360] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385304, 'name': CreateVM_Task, 'duration_secs': 0.331237} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 604.259950] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 604.260316] env[63202]: DEBUG oslo_concurrency.lockutils [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 604.260467] env[63202]: DEBUG oslo_concurrency.lockutils [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.260771] env[63202]: DEBUG oslo_concurrency.lockutils [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Acquired external semaphore "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 604.285230] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8fcd1f80-74f7-49fd-8be1-2bd93f80e6e7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.295276] env[63202]: DEBUG oslo_vmware.api [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Waiting for the task: (returnval){ [ 604.295276] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]5251e128-6c58-f7e6-17dc-0bc7039d8950" [ 604.295276] env[63202]: _type = "Task" [ 604.295276] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 604.301297] env[63202]: DEBUG nova.virt.hardware [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 604.301606] env[63202]: DEBUG nova.virt.hardware [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 604.301689] env[63202]: DEBUG nova.virt.hardware [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 604.301824] env[63202]: DEBUG nova.virt.hardware [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 604.301964] env[63202]: DEBUG nova.virt.hardware [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 604.302127] env[63202]: DEBUG nova.virt.hardware [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 604.302329] env[63202]: DEBUG nova.virt.hardware [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 604.302485] env[63202]: DEBUG nova.virt.hardware [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 604.302662] env[63202]: DEBUG nova.virt.hardware [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 604.302799] env[63202]: DEBUG nova.virt.hardware [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 604.302969] env[63202]: DEBUG nova.virt.hardware [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 604.303362] env[63202]: INFO nova.compute.manager [-] [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] Took 1.04 seconds to deallocate network for instance. [ 604.304107] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fadfcc8-face-424b-9d04-01d662073e7a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.308661] env[63202]: DEBUG nova.compute.claims [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 604.308869] env[63202]: DEBUG oslo_concurrency.lockutils [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.313889] env[63202]: DEBUG oslo_vmware.api [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]5251e128-6c58-f7e6-17dc-0bc7039d8950, 'name': SearchDatastore_Task, 'duration_secs': 0.009276} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 604.315985] env[63202]: DEBUG oslo_concurrency.lockutils [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 604.316223] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 604.316464] env[63202]: DEBUG oslo_concurrency.lockutils [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 604.316582] env[63202]: DEBUG oslo_concurrency.lockutils [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.316863] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 604.317144] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-76bc778e-0395-42be-8069-9bd6693cde98 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.319754] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49e83b0d-ef7e-4c94-b850-49f35f8d5f01 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.334461] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 604.334639] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Folder [datastore1] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 604.335415] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-347198a1-6993-479e-929b-f2357f2effaa {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.340103] env[63202]: DEBUG oslo_vmware.api [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Waiting for the task: (returnval){ [ 604.340103] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52bd0e57-af6e-a321-b275-89e5f13e030d" [ 604.340103] env[63202]: _type = "Task" [ 604.340103] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 604.348854] env[63202]: DEBUG oslo_vmware.api [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52bd0e57-af6e-a321-b275-89e5f13e030d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 604.457981] env[63202]: DEBUG oslo_concurrency.lockutils [req-0bd2a14b-9fc5-41fb-9802-294ab7216db9 req-f439376e-fefe-4bc9-bc1c-888e30005a8c service nova] Releasing lock "refresh_cache-90999eb1-8829-4f99-b3b9-8c70ee636cc5" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 604.530180] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.336s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 604.530771] env[63202]: DEBUG nova.compute.manager [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: c9b7182b-8116-419c-b399-2c7bcc133893] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 604.533564] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.340s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.537825] env[63202]: INFO nova.compute.claims [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 604.710468] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.854071] env[63202]: DEBUG oslo_vmware.api [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52bd0e57-af6e-a321-b275-89e5f13e030d, 'name': SearchDatastore_Task, 'duration_secs': 0.009189} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 604.855971] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-af8ebefd-6ab1-40ae-9f19-37e5b85e4971 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.864040] env[63202]: DEBUG oslo_vmware.api [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Waiting for the task: (returnval){ [ 604.864040] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52a63dfc-a535-e877-2f51-cce6f070a981" [ 604.864040] env[63202]: _type = "Task" [ 604.864040] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 604.874787] env[63202]: DEBUG oslo_vmware.api [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52a63dfc-a535-e877-2f51-cce6f070a981, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 605.046934] env[63202]: DEBUG nova.compute.utils [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 605.054204] env[63202]: DEBUG nova.compute.manager [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: c9b7182b-8116-419c-b399-2c7bcc133893] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 605.054204] env[63202]: DEBUG nova.network.neutron [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: c9b7182b-8116-419c-b399-2c7bcc133893] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 605.209090] env[63202]: DEBUG nova.policy [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '69493cd1ff694fe8a805f116b407ea1b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6b9be262a4894ae78a9964e5ca7b09a8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 605.264777] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] Acquiring lock "03e9f5cc-344c-45c8-abff-aae8db411832" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 605.265146] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] Lock "03e9f5cc-344c-45c8-abff-aae8db411832" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 605.374636] env[63202]: DEBUG oslo_vmware.api [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52a63dfc-a535-e877-2f51-cce6f070a981, 'name': SearchDatastore_Task, 'duration_secs': 0.010945} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 605.375315] env[63202]: DEBUG oslo_concurrency.lockutils [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.375674] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Copying Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] 0ded82aa-c68d-4d11-9e4d-90ed9a19b708/0ded82aa-c68d-4d11-9e4d-90ed9a19b708.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 605.376389] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8bec8e68-dbbd-459a-aa42-ef8aba43fc27 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.385031] env[63202]: DEBUG oslo_vmware.api [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Waiting for the task: (returnval){ [ 605.385031] env[63202]: value = "task-1385305" [ 605.385031] env[63202]: _type = "Task" [ 605.385031] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 605.391977] env[63202]: DEBUG oslo_vmware.api [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Task: {'id': task-1385305, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 605.559974] env[63202]: DEBUG nova.compute.manager [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: c9b7182b-8116-419c-b399-2c7bcc133893] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 605.804414] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4b5830d-6bf3-4b0f-a4c4-ff2b6ef1be62 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.813148] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-641db263-5b11-4076-b59a-9b2b6cbf1640 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.856964] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a01d1cf-0b0f-4537-9f2d-343923bc6eeb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.863764] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-230bead2-b0e6-4bf3-b9ed-464f58d9c280 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.886186] env[63202]: DEBUG nova.compute.provider_tree [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 605.897486] env[63202]: DEBUG oslo_vmware.api [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Task: {'id': task-1385305, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.482483} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 605.897678] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Copied Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] 0ded82aa-c68d-4d11-9e4d-90ed9a19b708/0ded82aa-c68d-4d11-9e4d-90ed9a19b708.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 605.897939] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 605.898344] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-152ef07d-ca6b-4649-b9b1-447df1065848 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.906147] env[63202]: DEBUG oslo_vmware.api [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Waiting for the task: (returnval){ [ 605.906147] env[63202]: value = "task-1385306" [ 605.906147] env[63202]: _type = "Task" [ 605.906147] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 605.914400] env[63202]: DEBUG oslo_vmware.api [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Task: {'id': task-1385306, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 606.082714] env[63202]: DEBUG nova.network.neutron [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: c9b7182b-8116-419c-b399-2c7bcc133893] Successfully created port: de336b0b-00ff-4033-bc80-b4b1d8e3fa67 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 606.265544] env[63202]: ERROR nova.compute.manager [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 684d1759-1f66-45f5-86c1-6fad8a310af1, please check neutron logs for more information. [ 606.265544] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 606.265544] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 606.265544] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 606.265544] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 606.265544] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 606.265544] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 606.265544] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 606.265544] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.265544] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 606.265544] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.265544] env[63202]: ERROR nova.compute.manager raise self.value [ 606.265544] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 606.265544] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 606.265544] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.265544] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 606.266248] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.266248] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 606.266248] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 684d1759-1f66-45f5-86c1-6fad8a310af1, please check neutron logs for more information. [ 606.266248] env[63202]: ERROR nova.compute.manager [ 606.266410] env[63202]: Traceback (most recent call last): [ 606.266410] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 606.266410] env[63202]: listener.cb(fileno) [ 606.266410] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 606.266410] env[63202]: result = function(*args, **kwargs) [ 606.266410] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 606.266410] env[63202]: return func(*args, **kwargs) [ 606.266410] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 606.266410] env[63202]: raise e [ 606.266410] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 606.266410] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 606.266410] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 606.266410] env[63202]: created_port_ids = self._update_ports_for_instance( [ 606.266410] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 606.266410] env[63202]: with excutils.save_and_reraise_exception(): [ 606.266410] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.266410] env[63202]: self.force_reraise() [ 606.266410] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.266410] env[63202]: raise self.value [ 606.266410] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 606.266410] env[63202]: updated_port = self._update_port( [ 606.266935] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.266935] env[63202]: _ensure_no_port_binding_failure(port) [ 606.266935] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.266935] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 606.266935] env[63202]: nova.exception.PortBindingFailed: Binding failed for port 684d1759-1f66-45f5-86c1-6fad8a310af1, please check neutron logs for more information. [ 606.266935] env[63202]: Removing descriptor: 16 [ 606.269164] env[63202]: ERROR nova.compute.manager [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 684d1759-1f66-45f5-86c1-6fad8a310af1, please check neutron logs for more information. [ 606.269164] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] Traceback (most recent call last): [ 606.269164] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 606.269164] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] yield resources [ 606.269164] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 606.269164] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] self.driver.spawn(context, instance, image_meta, [ 606.269164] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 606.269164] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 606.269164] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 606.269164] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] vm_ref = self.build_virtual_machine(instance, [ 606.269164] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 606.269664] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] vif_infos = vmwarevif.get_vif_info(self._session, [ 606.269664] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 606.269664] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] for vif in network_info: [ 606.269664] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 606.269664] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] return self._sync_wrapper(fn, *args, **kwargs) [ 606.269664] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 606.269664] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] self.wait() [ 606.269664] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 606.269664] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] self[:] = self._gt.wait() [ 606.269664] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 606.269664] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] return self._exit_event.wait() [ 606.269664] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 606.269664] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] result = hub.switch() [ 606.270017] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 606.270017] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] return self.greenlet.switch() [ 606.270017] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 606.270017] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] result = function(*args, **kwargs) [ 606.270017] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 606.270017] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] return func(*args, **kwargs) [ 606.270017] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 606.270017] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] raise e [ 606.270017] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 606.270017] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] nwinfo = self.network_api.allocate_for_instance( [ 606.270017] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 606.270017] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] created_port_ids = self._update_ports_for_instance( [ 606.270017] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 606.270341] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] with excutils.save_and_reraise_exception(): [ 606.270341] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.270341] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] self.force_reraise() [ 606.270341] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.270341] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] raise self.value [ 606.270341] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 606.270341] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] updated_port = self._update_port( [ 606.270341] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.270341] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] _ensure_no_port_binding_failure(port) [ 606.270341] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.270341] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] raise exception.PortBindingFailed(port_id=port['id']) [ 606.270341] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] nova.exception.PortBindingFailed: Binding failed for port 684d1759-1f66-45f5-86c1-6fad8a310af1, please check neutron logs for more information. [ 606.270341] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] [ 606.270986] env[63202]: INFO nova.compute.manager [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] Terminating instance [ 606.271817] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Acquiring lock "refresh_cache-c538fef4-2245-4bd3-af03-f0105fceb1e3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 606.272078] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Acquired lock "refresh_cache-c538fef4-2245-4bd3-af03-f0105fceb1e3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 606.272262] env[63202]: DEBUG nova.network.neutron [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 606.392965] env[63202]: DEBUG nova.scheduler.client.report [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 606.418097] env[63202]: DEBUG oslo_vmware.api [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Task: {'id': task-1385306, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.099482} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 606.418496] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 606.419335] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f94565c1-1158-484e-b3ed-5d1246838153 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.451375] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Reconfiguring VM instance instance-00000002 to attach disk [datastore1] 0ded82aa-c68d-4d11-9e4d-90ed9a19b708/0ded82aa-c68d-4d11-9e4d-90ed9a19b708.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 606.451375] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d14e8f6c-90be-421b-b266-cc42f6418578 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.476071] env[63202]: DEBUG oslo_vmware.api [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Waiting for the task: (returnval){ [ 606.476071] env[63202]: value = "task-1385307" [ 606.476071] env[63202]: _type = "Task" [ 606.476071] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 606.485267] env[63202]: DEBUG oslo_vmware.api [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Task: {'id': task-1385307, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 606.568820] env[63202]: DEBUG nova.compute.manager [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: c9b7182b-8116-419c-b399-2c7bcc133893] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 606.597167] env[63202]: DEBUG nova.virt.hardware [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 606.597421] env[63202]: DEBUG nova.virt.hardware [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 606.597529] env[63202]: DEBUG nova.virt.hardware [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 606.597705] env[63202]: DEBUG nova.virt.hardware [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 606.597848] env[63202]: DEBUG nova.virt.hardware [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 606.597990] env[63202]: DEBUG nova.virt.hardware [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 606.598550] env[63202]: DEBUG nova.virt.hardware [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 606.598766] env[63202]: DEBUG nova.virt.hardware [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 606.598880] env[63202]: DEBUG nova.virt.hardware [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 606.599054] env[63202]: DEBUG nova.virt.hardware [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 606.599252] env[63202]: DEBUG nova.virt.hardware [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 606.600290] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-360f6654-ae36-4e7c-96cc-73cd5a8c50c2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.608358] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ac1af3a-31e7-415c-ab03-a3c0c5ccf1ed {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.900142] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.366s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 606.900661] env[63202]: DEBUG nova.compute.manager [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 606.903375] env[63202]: DEBUG oslo_concurrency.lockutils [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.031s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.904972] env[63202]: INFO nova.compute.claims [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 606.939228] env[63202]: DEBUG nova.network.neutron [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 606.992242] env[63202]: DEBUG oslo_vmware.api [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Task: {'id': task-1385307, 'name': ReconfigVM_Task, 'duration_secs': 0.27587} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 606.992521] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Reconfigured VM instance instance-00000002 to attach disk [datastore1] 0ded82aa-c68d-4d11-9e4d-90ed9a19b708/0ded82aa-c68d-4d11-9e4d-90ed9a19b708.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 606.994258] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-13bcb593-b286-4124-96cc-0114cf5277ea {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.001458] env[63202]: DEBUG oslo_vmware.api [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Waiting for the task: (returnval){ [ 607.001458] env[63202]: value = "task-1385308" [ 607.001458] env[63202]: _type = "Task" [ 607.001458] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 607.016513] env[63202]: DEBUG oslo_vmware.api [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Task: {'id': task-1385308, 'name': Rename_Task} progress is 6%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 607.321780] env[63202]: DEBUG nova.network.neutron [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.409125] env[63202]: DEBUG nova.compute.utils [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 607.415994] env[63202]: DEBUG nova.compute.manager [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 607.416244] env[63202]: DEBUG nova.network.neutron [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 607.513259] env[63202]: DEBUG oslo_vmware.api [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Task: {'id': task-1385308, 'name': Rename_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 607.643233] env[63202]: DEBUG nova.policy [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '69493cd1ff694fe8a805f116b407ea1b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6b9be262a4894ae78a9964e5ca7b09a8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 607.826367] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Releasing lock "refresh_cache-c538fef4-2245-4bd3-af03-f0105fceb1e3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 607.826620] env[63202]: DEBUG nova.compute.manager [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 607.826875] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 607.827100] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-499bbab6-0696-478d-9443-912cab4ae381 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.836428] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88a8f3b6-7d7f-4f99-b041-220cb13688f9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.862297] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c538fef4-2245-4bd3-af03-f0105fceb1e3 could not be found. [ 607.862573] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 607.862816] env[63202]: INFO nova.compute.manager [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] Took 0.04 seconds to destroy the instance on the hypervisor. [ 607.863021] env[63202]: DEBUG oslo.service.loopingcall [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 607.863249] env[63202]: DEBUG nova.compute.manager [-] [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 607.863360] env[63202]: DEBUG nova.network.neutron [-] [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 607.926105] env[63202]: DEBUG nova.compute.manager [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 608.018931] env[63202]: DEBUG oslo_vmware.api [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Task: {'id': task-1385308, 'name': Rename_Task, 'duration_secs': 0.825263} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 608.019427] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 608.019708] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4fae7b2e-cf08-4fa7-be56-af7e5e570f89 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.028877] env[63202]: DEBUG oslo_vmware.api [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Waiting for the task: (returnval){ [ 608.028877] env[63202]: value = "task-1385309" [ 608.028877] env[63202]: _type = "Task" [ 608.028877] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 608.038348] env[63202]: DEBUG oslo_vmware.api [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Task: {'id': task-1385309, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 608.187031] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa1553ac-1959-4f4c-9abd-445b129442f1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.195015] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c444c23-c7ca-4266-a8b0-9ff0aeb3aa05 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.228831] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c199d13-747e-4c11-a2f6-3d4b5db9abe1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.238333] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f510490-9a6c-4734-9107-f80f6159b9cd {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.253739] env[63202]: DEBUG nova.compute.provider_tree [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 608.285249] env[63202]: ERROR nova.compute.manager [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b35498a4-9923-43db-8dda-1e3802f5777a, please check neutron logs for more information. [ 608.285249] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 608.285249] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 608.285249] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 608.285249] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 608.285249] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 608.285249] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 608.285249] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 608.285249] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 608.285249] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 608.285249] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 608.285249] env[63202]: ERROR nova.compute.manager raise self.value [ 608.285249] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 608.285249] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 608.285249] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 608.285249] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 608.285884] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 608.285884] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 608.285884] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b35498a4-9923-43db-8dda-1e3802f5777a, please check neutron logs for more information. [ 608.285884] env[63202]: ERROR nova.compute.manager [ 608.285884] env[63202]: Traceback (most recent call last): [ 608.285884] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 608.285884] env[63202]: listener.cb(fileno) [ 608.285884] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 608.285884] env[63202]: result = function(*args, **kwargs) [ 608.285884] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 608.285884] env[63202]: return func(*args, **kwargs) [ 608.285884] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 608.285884] env[63202]: raise e [ 608.285884] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 608.285884] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 608.285884] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 608.285884] env[63202]: created_port_ids = self._update_ports_for_instance( [ 608.285884] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 608.285884] env[63202]: with excutils.save_and_reraise_exception(): [ 608.285884] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 608.285884] env[63202]: self.force_reraise() [ 608.285884] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 608.285884] env[63202]: raise self.value [ 608.285884] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 608.285884] env[63202]: updated_port = self._update_port( [ 608.285884] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 608.285884] env[63202]: _ensure_no_port_binding_failure(port) [ 608.285884] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 608.285884] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 608.287259] env[63202]: nova.exception.PortBindingFailed: Binding failed for port b35498a4-9923-43db-8dda-1e3802f5777a, please check neutron logs for more information. [ 608.287259] env[63202]: Removing descriptor: 17 [ 608.287259] env[63202]: ERROR nova.compute.manager [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b35498a4-9923-43db-8dda-1e3802f5777a, please check neutron logs for more information. [ 608.287259] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] Traceback (most recent call last): [ 608.287259] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 608.287259] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] yield resources [ 608.287259] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 608.287259] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] self.driver.spawn(context, instance, image_meta, [ 608.287259] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 608.287259] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] self._vmops.spawn(context, instance, image_meta, injected_files, [ 608.287259] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 608.287259] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] vm_ref = self.build_virtual_machine(instance, [ 608.287955] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 608.287955] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] vif_infos = vmwarevif.get_vif_info(self._session, [ 608.287955] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 608.287955] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] for vif in network_info: [ 608.287955] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 608.287955] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] return self._sync_wrapper(fn, *args, **kwargs) [ 608.287955] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 608.287955] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] self.wait() [ 608.287955] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 608.287955] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] self[:] = self._gt.wait() [ 608.287955] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 608.287955] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] return self._exit_event.wait() [ 608.287955] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 608.289163] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] result = hub.switch() [ 608.289163] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 608.289163] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] return self.greenlet.switch() [ 608.289163] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 608.289163] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] result = function(*args, **kwargs) [ 608.289163] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 608.289163] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] return func(*args, **kwargs) [ 608.289163] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 608.289163] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] raise e [ 608.289163] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 608.289163] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] nwinfo = self.network_api.allocate_for_instance( [ 608.289163] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 608.289163] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] created_port_ids = self._update_ports_for_instance( [ 608.289526] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 608.289526] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] with excutils.save_and_reraise_exception(): [ 608.289526] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 608.289526] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] self.force_reraise() [ 608.289526] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 608.289526] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] raise self.value [ 608.289526] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 608.289526] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] updated_port = self._update_port( [ 608.289526] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 608.289526] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] _ensure_no_port_binding_failure(port) [ 608.289526] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 608.289526] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] raise exception.PortBindingFailed(port_id=port['id']) [ 608.289932] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] nova.exception.PortBindingFailed: Binding failed for port b35498a4-9923-43db-8dda-1e3802f5777a, please check neutron logs for more information. [ 608.289932] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] [ 608.289932] env[63202]: INFO nova.compute.manager [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] Terminating instance [ 608.289932] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] Acquiring lock "refresh_cache-2f757f9b-537e-4bd8-b08d-1da991903552" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 608.289932] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] Acquired lock "refresh_cache-2f757f9b-537e-4bd8-b08d-1da991903552" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 608.289932] env[63202]: DEBUG nova.network.neutron [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 608.366518] env[63202]: DEBUG nova.network.neutron [-] [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 608.464601] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3a242ef3-a939-47fb-8baa-b578de6860c3 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Acquiring lock "21d0b283-3bb4-4bda-8e62-c933c6de4927" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 608.464761] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3a242ef3-a939-47fb-8baa-b578de6860c3 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Lock "21d0b283-3bb4-4bda-8e62-c933c6de4927" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 608.464984] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3a242ef3-a939-47fb-8baa-b578de6860c3 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Acquiring lock "21d0b283-3bb4-4bda-8e62-c933c6de4927-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 608.465181] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3a242ef3-a939-47fb-8baa-b578de6860c3 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Lock "21d0b283-3bb4-4bda-8e62-c933c6de4927-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 608.465383] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3a242ef3-a939-47fb-8baa-b578de6860c3 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Lock "21d0b283-3bb4-4bda-8e62-c933c6de4927-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 608.468045] env[63202]: INFO nova.compute.manager [None req-3a242ef3-a939-47fb-8baa-b578de6860c3 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] [instance: 21d0b283-3bb4-4bda-8e62-c933c6de4927] Terminating instance [ 608.470371] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3a242ef3-a939-47fb-8baa-b578de6860c3 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Acquiring lock "refresh_cache-21d0b283-3bb4-4bda-8e62-c933c6de4927" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 608.470771] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3a242ef3-a939-47fb-8baa-b578de6860c3 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Acquired lock "refresh_cache-21d0b283-3bb4-4bda-8e62-c933c6de4927" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 608.470771] env[63202]: DEBUG nova.network.neutron [None req-3a242ef3-a939-47fb-8baa-b578de6860c3 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] [instance: 21d0b283-3bb4-4bda-8e62-c933c6de4927] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 608.542023] env[63202]: DEBUG oslo_vmware.api [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Task: {'id': task-1385309, 'name': PowerOnVM_Task, 'duration_secs': 0.450684} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 608.542516] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 608.543016] env[63202]: DEBUG nova.compute.manager [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 608.543700] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a21dbd04-a64d-4b57-9fbb-0af0db20f7f2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.663777] env[63202]: DEBUG nova.compute.manager [req-b8132749-ba16-45cc-8f86-948a40963158 req-e9dd755c-b8de-4866-be6a-bf4ddaeec915 service nova] [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] Received event network-vif-deleted-180d9c18-47ea-4305-a5b8-ab5abcdf5a1a {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 608.756938] env[63202]: DEBUG nova.scheduler.client.report [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 608.859674] env[63202]: DEBUG nova.network.neutron [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 608.864697] env[63202]: DEBUG nova.network.neutron [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] Successfully created port: fdd4ea5c-13a8-48f2-b391-b5e0c0e7241a {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 608.939514] env[63202]: DEBUG nova.compute.manager [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 608.968513] env[63202]: DEBUG nova.virt.hardware [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 608.968713] env[63202]: DEBUG nova.virt.hardware [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 608.968862] env[63202]: DEBUG nova.virt.hardware [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 608.969051] env[63202]: DEBUG nova.virt.hardware [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 608.969197] env[63202]: DEBUG nova.virt.hardware [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 608.969340] env[63202]: DEBUG nova.virt.hardware [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 608.969674] env[63202]: DEBUG nova.virt.hardware [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 608.969747] env[63202]: DEBUG nova.virt.hardware [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 608.969914] env[63202]: DEBUG nova.virt.hardware [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 608.971882] env[63202]: DEBUG nova.virt.hardware [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 608.971882] env[63202]: DEBUG nova.virt.hardware [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 608.972844] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a137384f-88d5-40fe-bb5d-1b7f972842fc {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.988998] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee01aecc-61c7-4b75-a8ff-5965bbe5ef47 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.037040] env[63202]: DEBUG nova.network.neutron [None req-3a242ef3-a939-47fb-8baa-b578de6860c3 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] [instance: 21d0b283-3bb4-4bda-8e62-c933c6de4927] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 609.068656] env[63202]: DEBUG oslo_concurrency.lockutils [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 609.183027] env[63202]: DEBUG nova.network.neutron [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.265861] env[63202]: DEBUG oslo_concurrency.lockutils [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.363s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 609.266711] env[63202]: DEBUG nova.compute.manager [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 609.272032] env[63202]: DEBUG oslo_concurrency.lockutils [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.272s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 609.355191] env[63202]: DEBUG nova.network.neutron [None req-3a242ef3-a939-47fb-8baa-b578de6860c3 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] [instance: 21d0b283-3bb4-4bda-8e62-c933c6de4927] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.690476] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] Releasing lock "refresh_cache-2f757f9b-537e-4bd8-b08d-1da991903552" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 609.690476] env[63202]: DEBUG nova.compute.manager [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 609.690476] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 609.690476] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-aa656367-cc6c-4464-b575-cc65c3184008 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.703040] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c2b74bc-8414-4a42-9479-0c5e256c6b92 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.729934] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2f757f9b-537e-4bd8-b08d-1da991903552 could not be found. [ 609.730194] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 609.730380] env[63202]: INFO nova.compute.manager [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] Took 0.04 seconds to destroy the instance on the hypervisor. [ 609.730630] env[63202]: DEBUG oslo.service.loopingcall [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 609.730859] env[63202]: DEBUG nova.compute.manager [-] [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 609.730951] env[63202]: DEBUG nova.network.neutron [-] [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 609.773240] env[63202]: DEBUG nova.network.neutron [-] [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 609.782015] env[63202]: DEBUG nova.compute.utils [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 609.789337] env[63202]: DEBUG nova.compute.manager [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 609.790084] env[63202]: DEBUG nova.network.neutron [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 609.857351] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3a242ef3-a939-47fb-8baa-b578de6860c3 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Releasing lock "refresh_cache-21d0b283-3bb4-4bda-8e62-c933c6de4927" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 609.858040] env[63202]: DEBUG nova.compute.manager [None req-3a242ef3-a939-47fb-8baa-b578de6860c3 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] [instance: 21d0b283-3bb4-4bda-8e62-c933c6de4927] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 609.858282] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-3a242ef3-a939-47fb-8baa-b578de6860c3 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] [instance: 21d0b283-3bb4-4bda-8e62-c933c6de4927] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 609.859153] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73e62749-4a04-4d5b-9d34-2b7dfd971f76 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.868070] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a242ef3-a939-47fb-8baa-b578de6860c3 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] [instance: 21d0b283-3bb4-4bda-8e62-c933c6de4927] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 609.871145] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9ce8b68d-3656-46d8-971e-6544494b0619 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.878035] env[63202]: DEBUG oslo_vmware.api [None req-3a242ef3-a939-47fb-8baa-b578de6860c3 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Waiting for the task: (returnval){ [ 609.878035] env[63202]: value = "task-1385310" [ 609.878035] env[63202]: _type = "Task" [ 609.878035] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 609.894125] env[63202]: DEBUG oslo_vmware.api [None req-3a242ef3-a939-47fb-8baa-b578de6860c3 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Task: {'id': task-1385310, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.069438] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf78e2c4-07a0-4d50-acc6-f29045eaad28 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.078185] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc705218-c1b7-4c35-b939-b7f7cc6e534a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.115572] env[63202]: ERROR nova.compute.manager [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port de336b0b-00ff-4033-bc80-b4b1d8e3fa67, please check neutron logs for more information. [ 610.115572] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 610.115572] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.115572] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 610.115572] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 610.115572] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 610.115572] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 610.115572] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 610.115572] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.115572] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 610.115572] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.115572] env[63202]: ERROR nova.compute.manager raise self.value [ 610.115572] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 610.115572] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 610.115572] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.115572] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 610.116086] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.116086] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 610.116086] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port de336b0b-00ff-4033-bc80-b4b1d8e3fa67, please check neutron logs for more information. [ 610.116086] env[63202]: ERROR nova.compute.manager [ 610.116086] env[63202]: Traceback (most recent call last): [ 610.116086] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 610.116086] env[63202]: listener.cb(fileno) [ 610.116086] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 610.116086] env[63202]: result = function(*args, **kwargs) [ 610.116086] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 610.116086] env[63202]: return func(*args, **kwargs) [ 610.116086] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 610.116086] env[63202]: raise e [ 610.116086] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.116086] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 610.116086] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 610.116086] env[63202]: created_port_ids = self._update_ports_for_instance( [ 610.116086] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 610.116086] env[63202]: with excutils.save_and_reraise_exception(): [ 610.116086] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.116086] env[63202]: self.force_reraise() [ 610.116086] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.116086] env[63202]: raise self.value [ 610.116086] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 610.116086] env[63202]: updated_port = self._update_port( [ 610.116086] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.116086] env[63202]: _ensure_no_port_binding_failure(port) [ 610.116086] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.116086] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 610.116854] env[63202]: nova.exception.PortBindingFailed: Binding failed for port de336b0b-00ff-4033-bc80-b4b1d8e3fa67, please check neutron logs for more information. [ 610.116854] env[63202]: Removing descriptor: 14 [ 610.116854] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a637795c-57e6-46e8-8556-66b992212ff4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.119852] env[63202]: ERROR nova.compute.manager [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: c9b7182b-8116-419c-b399-2c7bcc133893] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port de336b0b-00ff-4033-bc80-b4b1d8e3fa67, please check neutron logs for more information. [ 610.119852] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] Traceback (most recent call last): [ 610.119852] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 610.119852] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] yield resources [ 610.119852] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 610.119852] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] self.driver.spawn(context, instance, image_meta, [ 610.119852] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 610.119852] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] self._vmops.spawn(context, instance, image_meta, injected_files, [ 610.119852] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 610.119852] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] vm_ref = self.build_virtual_machine(instance, [ 610.119852] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 610.120485] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] vif_infos = vmwarevif.get_vif_info(self._session, [ 610.120485] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 610.120485] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] for vif in network_info: [ 610.120485] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 610.120485] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] return self._sync_wrapper(fn, *args, **kwargs) [ 610.120485] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 610.120485] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] self.wait() [ 610.120485] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 610.120485] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] self[:] = self._gt.wait() [ 610.120485] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 610.120485] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] return self._exit_event.wait() [ 610.120485] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 610.120485] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] result = hub.switch() [ 610.122083] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 610.122083] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] return self.greenlet.switch() [ 610.122083] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 610.122083] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] result = function(*args, **kwargs) [ 610.122083] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 610.122083] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] return func(*args, **kwargs) [ 610.122083] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 610.122083] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] raise e [ 610.122083] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.122083] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] nwinfo = self.network_api.allocate_for_instance( [ 610.122083] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 610.122083] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] created_port_ids = self._update_ports_for_instance( [ 610.122083] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 610.122794] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] with excutils.save_and_reraise_exception(): [ 610.122794] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.122794] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] self.force_reraise() [ 610.122794] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.122794] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] raise self.value [ 610.122794] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 610.122794] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] updated_port = self._update_port( [ 610.122794] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.122794] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] _ensure_no_port_binding_failure(port) [ 610.122794] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.122794] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] raise exception.PortBindingFailed(port_id=port['id']) [ 610.122794] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] nova.exception.PortBindingFailed: Binding failed for port de336b0b-00ff-4033-bc80-b4b1d8e3fa67, please check neutron logs for more information. [ 610.122794] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] [ 610.123594] env[63202]: INFO nova.compute.manager [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: c9b7182b-8116-419c-b399-2c7bcc133893] Terminating instance [ 610.123594] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Acquiring lock "refresh_cache-c9b7182b-8116-419c-b399-2c7bcc133893" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 610.123594] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Acquired lock "refresh_cache-c9b7182b-8116-419c-b399-2c7bcc133893" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.123594] env[63202]: DEBUG nova.network.neutron [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: c9b7182b-8116-419c-b399-2c7bcc133893] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 610.127893] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f62c467-67c2-4bfe-be88-f681aa265c5a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.142065] env[63202]: DEBUG nova.compute.provider_tree [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 610.156033] env[63202]: DEBUG nova.policy [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2bd91973e1bb4ce7be8bce9bfe34a941', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e972faeaedc6468aab7e7cfee88a477b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 610.264275] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] Acquiring lock "ef69de02-6e43-4489-a35a-8a9096e54fbd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.264562] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] Lock "ef69de02-6e43-4489-a35a-8a9096e54fbd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.279593] env[63202]: DEBUG nova.network.neutron [-] [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.288926] env[63202]: DEBUG nova.compute.manager [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 610.390611] env[63202]: DEBUG oslo_vmware.api [None req-3a242ef3-a939-47fb-8baa-b578de6860c3 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Task: {'id': task-1385310, 'name': PowerOffVM_Task, 'duration_secs': 0.256301} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 610.391249] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a242ef3-a939-47fb-8baa-b578de6860c3 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] [instance: 21d0b283-3bb4-4bda-8e62-c933c6de4927] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 610.391488] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-3a242ef3-a939-47fb-8baa-b578de6860c3 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] [instance: 21d0b283-3bb4-4bda-8e62-c933c6de4927] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 610.391746] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-61e9f75b-a26c-4dd8-92b7-604cec2387b3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.420371] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-3a242ef3-a939-47fb-8baa-b578de6860c3 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] [instance: 21d0b283-3bb4-4bda-8e62-c933c6de4927] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 610.420646] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-3a242ef3-a939-47fb-8baa-b578de6860c3 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] [instance: 21d0b283-3bb4-4bda-8e62-c933c6de4927] Deleting contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 610.420863] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a242ef3-a939-47fb-8baa-b578de6860c3 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Deleting the datastore file [datastore1] 21d0b283-3bb4-4bda-8e62-c933c6de4927 {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 610.421152] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cc1bc6fc-7649-4a67-9ace-44462cc17116 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.428528] env[63202]: DEBUG oslo_vmware.api [None req-3a242ef3-a939-47fb-8baa-b578de6860c3 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Waiting for the task: (returnval){ [ 610.428528] env[63202]: value = "task-1385312" [ 610.428528] env[63202]: _type = "Task" [ 610.428528] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 610.437271] env[63202]: DEBUG oslo_vmware.api [None req-3a242ef3-a939-47fb-8baa-b578de6860c3 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Task: {'id': task-1385312, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.645755] env[63202]: DEBUG nova.scheduler.client.report [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 610.683819] env[63202]: DEBUG nova.network.neutron [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: c9b7182b-8116-419c-b399-2c7bcc133893] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 610.692922] env[63202]: DEBUG nova.network.neutron [-] [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.785353] env[63202]: INFO nova.compute.manager [-] [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] Took 1.05 seconds to deallocate network for instance. [ 610.788049] env[63202]: DEBUG nova.compute.claims [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 610.788262] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.855650] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Acquiring lock "bc8fc812-1e66-4e14-a39c-1c226e7fe2b9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.855844] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Lock "bc8fc812-1e66-4e14-a39c-1c226e7fe2b9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.940497] env[63202]: DEBUG oslo_vmware.api [None req-3a242ef3-a939-47fb-8baa-b578de6860c3 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Task: {'id': task-1385312, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.099822} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 610.940497] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a242ef3-a939-47fb-8baa-b578de6860c3 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 610.940497] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-3a242ef3-a939-47fb-8baa-b578de6860c3 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] [instance: 21d0b283-3bb4-4bda-8e62-c933c6de4927] Deleted contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 610.940497] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-3a242ef3-a939-47fb-8baa-b578de6860c3 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] [instance: 21d0b283-3bb4-4bda-8e62-c933c6de4927] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 610.940641] env[63202]: INFO nova.compute.manager [None req-3a242ef3-a939-47fb-8baa-b578de6860c3 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] [instance: 21d0b283-3bb4-4bda-8e62-c933c6de4927] Took 1.08 seconds to destroy the instance on the hypervisor. [ 610.941056] env[63202]: DEBUG oslo.service.loopingcall [None req-3a242ef3-a939-47fb-8baa-b578de6860c3 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 610.941741] env[63202]: DEBUG nova.compute.manager [-] [instance: 21d0b283-3bb4-4bda-8e62-c933c6de4927] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 610.941741] env[63202]: DEBUG nova.network.neutron [-] [instance: 21d0b283-3bb4-4bda-8e62-c933c6de4927] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 610.993855] env[63202]: DEBUG nova.network.neutron [-] [instance: 21d0b283-3bb4-4bda-8e62-c933c6de4927] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 611.063056] env[63202]: DEBUG nova.network.neutron [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: c9b7182b-8116-419c-b399-2c7bcc133893] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.150941] env[63202]: DEBUG oslo_concurrency.lockutils [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.881s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 611.152312] env[63202]: ERROR nova.compute.manager [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5305bbd8-e698-4cb8-a36d-de3f0148ecac, please check neutron logs for more information. [ 611.152312] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] Traceback (most recent call last): [ 611.152312] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 611.152312] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] self.driver.spawn(context, instance, image_meta, [ 611.152312] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 611.152312] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 611.152312] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 611.152312] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] vm_ref = self.build_virtual_machine(instance, [ 611.152312] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 611.152312] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] vif_infos = vmwarevif.get_vif_info(self._session, [ 611.152312] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 611.152758] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] for vif in network_info: [ 611.152758] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 611.152758] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] return self._sync_wrapper(fn, *args, **kwargs) [ 611.152758] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 611.152758] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] self.wait() [ 611.152758] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 611.152758] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] self[:] = self._gt.wait() [ 611.152758] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 611.152758] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] return self._exit_event.wait() [ 611.152758] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 611.152758] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] result = hub.switch() [ 611.152758] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 611.152758] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] return self.greenlet.switch() [ 611.153239] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 611.153239] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] result = function(*args, **kwargs) [ 611.153239] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 611.153239] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] return func(*args, **kwargs) [ 611.153239] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 611.153239] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] raise e [ 611.153239] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.153239] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] nwinfo = self.network_api.allocate_for_instance( [ 611.153239] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 611.153239] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] created_port_ids = self._update_ports_for_instance( [ 611.153239] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 611.153239] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] with excutils.save_and_reraise_exception(): [ 611.153239] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.153793] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] self.force_reraise() [ 611.153793] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.153793] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] raise self.value [ 611.153793] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 611.153793] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] updated_port = self._update_port( [ 611.153793] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.153793] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] _ensure_no_port_binding_failure(port) [ 611.153793] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.153793] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] raise exception.PortBindingFailed(port_id=port['id']) [ 611.153793] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] nova.exception.PortBindingFailed: Binding failed for port 5305bbd8-e698-4cb8-a36d-de3f0148ecac, please check neutron logs for more information. [ 611.153793] env[63202]: ERROR nova.compute.manager [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] [ 611.154729] env[63202]: DEBUG nova.compute.utils [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] Binding failed for port 5305bbd8-e698-4cb8-a36d-de3f0148ecac, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 611.156569] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.875s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.158678] env[63202]: INFO nova.compute.claims [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 611.163355] env[63202]: DEBUG nova.compute.manager [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] Build of instance 6e102794-152e-492e-b063-e1ebaffa82b6 was re-scheduled: Binding failed for port 5305bbd8-e698-4cb8-a36d-de3f0148ecac, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 611.163797] env[63202]: DEBUG nova.compute.manager [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 611.163880] env[63202]: DEBUG oslo_concurrency.lockutils [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] Acquiring lock "refresh_cache-6e102794-152e-492e-b063-e1ebaffa82b6" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.164038] env[63202]: DEBUG oslo_concurrency.lockutils [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] Acquired lock "refresh_cache-6e102794-152e-492e-b063-e1ebaffa82b6" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.164271] env[63202]: DEBUG nova.network.neutron [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 611.198094] env[63202]: INFO nova.compute.manager [-] [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] Took 3.33 seconds to deallocate network for instance. [ 611.199310] env[63202]: DEBUG nova.compute.claims [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 611.199873] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.217853] env[63202]: ERROR nova.compute.manager [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fdd4ea5c-13a8-48f2-b391-b5e0c0e7241a, please check neutron logs for more information. [ 611.217853] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 611.217853] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.217853] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 611.217853] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 611.217853] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 611.217853] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 611.217853] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 611.217853] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.217853] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 611.217853] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.217853] env[63202]: ERROR nova.compute.manager raise self.value [ 611.217853] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 611.217853] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 611.217853] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.217853] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 611.219878] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.219878] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 611.219878] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fdd4ea5c-13a8-48f2-b391-b5e0c0e7241a, please check neutron logs for more information. [ 611.219878] env[63202]: ERROR nova.compute.manager [ 611.219878] env[63202]: Traceback (most recent call last): [ 611.219878] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 611.219878] env[63202]: listener.cb(fileno) [ 611.219878] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 611.219878] env[63202]: result = function(*args, **kwargs) [ 611.219878] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 611.219878] env[63202]: return func(*args, **kwargs) [ 611.219878] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 611.219878] env[63202]: raise e [ 611.219878] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.219878] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 611.219878] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 611.219878] env[63202]: created_port_ids = self._update_ports_for_instance( [ 611.219878] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 611.219878] env[63202]: with excutils.save_and_reraise_exception(): [ 611.219878] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.219878] env[63202]: self.force_reraise() [ 611.219878] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.219878] env[63202]: raise self.value [ 611.219878] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 611.219878] env[63202]: updated_port = self._update_port( [ 611.219878] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.219878] env[63202]: _ensure_no_port_binding_failure(port) [ 611.219878] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.219878] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 611.220848] env[63202]: nova.exception.PortBindingFailed: Binding failed for port fdd4ea5c-13a8-48f2-b391-b5e0c0e7241a, please check neutron logs for more information. [ 611.220848] env[63202]: Removing descriptor: 16 [ 611.220848] env[63202]: ERROR nova.compute.manager [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fdd4ea5c-13a8-48f2-b391-b5e0c0e7241a, please check neutron logs for more information. [ 611.220848] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] Traceback (most recent call last): [ 611.220848] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 611.220848] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] yield resources [ 611.220848] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 611.220848] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] self.driver.spawn(context, instance, image_meta, [ 611.220848] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 611.220848] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 611.220848] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 611.220848] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] vm_ref = self.build_virtual_machine(instance, [ 611.221231] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 611.221231] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] vif_infos = vmwarevif.get_vif_info(self._session, [ 611.221231] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 611.221231] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] for vif in network_info: [ 611.221231] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 611.221231] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] return self._sync_wrapper(fn, *args, **kwargs) [ 611.221231] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 611.221231] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] self.wait() [ 611.221231] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 611.221231] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] self[:] = self._gt.wait() [ 611.221231] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 611.221231] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] return self._exit_event.wait() [ 611.221231] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 611.221636] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] result = hub.switch() [ 611.221636] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 611.221636] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] return self.greenlet.switch() [ 611.221636] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 611.221636] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] result = function(*args, **kwargs) [ 611.221636] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 611.221636] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] return func(*args, **kwargs) [ 611.221636] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 611.221636] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] raise e [ 611.221636] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.221636] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] nwinfo = self.network_api.allocate_for_instance( [ 611.221636] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 611.221636] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] created_port_ids = self._update_ports_for_instance( [ 611.222051] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 611.222051] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] with excutils.save_and_reraise_exception(): [ 611.222051] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.222051] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] self.force_reraise() [ 611.222051] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.222051] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] raise self.value [ 611.222051] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 611.222051] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] updated_port = self._update_port( [ 611.222051] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.222051] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] _ensure_no_port_binding_failure(port) [ 611.222051] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.222051] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] raise exception.PortBindingFailed(port_id=port['id']) [ 611.222393] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] nova.exception.PortBindingFailed: Binding failed for port fdd4ea5c-13a8-48f2-b391-b5e0c0e7241a, please check neutron logs for more information. [ 611.222393] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] [ 611.222393] env[63202]: INFO nova.compute.manager [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] Terminating instance [ 611.224226] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Acquiring lock "refresh_cache-4f0aba5d-1722-4ddb-8331-a10c25e29b6f" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.224226] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Acquired lock "refresh_cache-4f0aba5d-1722-4ddb-8331-a10c25e29b6f" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.224226] env[63202]: DEBUG nova.network.neutron [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 611.300471] env[63202]: DEBUG nova.compute.manager [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 611.338172] env[63202]: DEBUG nova.network.neutron [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] Successfully created port: f47aceff-c181-4583-8852-323723733b26 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 611.339954] env[63202]: DEBUG nova.virt.hardware [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 611.340397] env[63202]: DEBUG nova.virt.hardware [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 611.340671] env[63202]: DEBUG nova.virt.hardware [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 611.341074] env[63202]: DEBUG nova.virt.hardware [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 611.341420] env[63202]: DEBUG nova.virt.hardware [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 611.341667] env[63202]: DEBUG nova.virt.hardware [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 611.341973] env[63202]: DEBUG nova.virt.hardware [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 611.342252] env[63202]: DEBUG nova.virt.hardware [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 611.342541] env[63202]: DEBUG nova.virt.hardware [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 611.342898] env[63202]: DEBUG nova.virt.hardware [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 611.343186] env[63202]: DEBUG nova.virt.hardware [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 611.344484] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ad33df3-bbdc-444c-b1f0-14d72dd3db6b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.356308] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15d5f8e8-4807-46b9-9f49-96562f61cfd1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.426820] env[63202]: DEBUG nova.compute.manager [req-efd3fa99-f0b3-4bde-a9a9-2c16601f2415 req-b81eb92f-d1ae-407f-84bb-7fb5f9c83001 service nova] [instance: c9b7182b-8116-419c-b399-2c7bcc133893] Received event network-changed-de336b0b-00ff-4033-bc80-b4b1d8e3fa67 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 611.427705] env[63202]: DEBUG nova.compute.manager [req-efd3fa99-f0b3-4bde-a9a9-2c16601f2415 req-b81eb92f-d1ae-407f-84bb-7fb5f9c83001 service nova] [instance: c9b7182b-8116-419c-b399-2c7bcc133893] Refreshing instance network info cache due to event network-changed-de336b0b-00ff-4033-bc80-b4b1d8e3fa67. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 611.427705] env[63202]: DEBUG oslo_concurrency.lockutils [req-efd3fa99-f0b3-4bde-a9a9-2c16601f2415 req-b81eb92f-d1ae-407f-84bb-7fb5f9c83001 service nova] Acquiring lock "refresh_cache-c9b7182b-8116-419c-b399-2c7bcc133893" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.501424] env[63202]: DEBUG nova.network.neutron [-] [instance: 21d0b283-3bb4-4bda-8e62-c933c6de4927] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.568763] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Releasing lock "refresh_cache-c9b7182b-8116-419c-b399-2c7bcc133893" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 611.569111] env[63202]: DEBUG nova.compute.manager [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: c9b7182b-8116-419c-b399-2c7bcc133893] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 611.569308] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: c9b7182b-8116-419c-b399-2c7bcc133893] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 611.569719] env[63202]: DEBUG oslo_concurrency.lockutils [req-efd3fa99-f0b3-4bde-a9a9-2c16601f2415 req-b81eb92f-d1ae-407f-84bb-7fb5f9c83001 service nova] Acquired lock "refresh_cache-c9b7182b-8116-419c-b399-2c7bcc133893" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.569792] env[63202]: DEBUG nova.network.neutron [req-efd3fa99-f0b3-4bde-a9a9-2c16601f2415 req-b81eb92f-d1ae-407f-84bb-7fb5f9c83001 service nova] [instance: c9b7182b-8116-419c-b399-2c7bcc133893] Refreshing network info cache for port de336b0b-00ff-4033-bc80-b4b1d8e3fa67 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 611.570834] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f1ddca38-4613-4c04-95ae-ea3d47023ecf {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.585224] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-132dee41-84da-4544-8366-40493151e6ed {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.612317] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: c9b7182b-8116-419c-b399-2c7bcc133893] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c9b7182b-8116-419c-b399-2c7bcc133893 could not be found. [ 611.612564] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: c9b7182b-8116-419c-b399-2c7bcc133893] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 611.613128] env[63202]: INFO nova.compute.manager [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: c9b7182b-8116-419c-b399-2c7bcc133893] Took 0.04 seconds to destroy the instance on the hypervisor. [ 611.613128] env[63202]: DEBUG oslo.service.loopingcall [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 611.613243] env[63202]: DEBUG nova.compute.manager [-] [instance: c9b7182b-8116-419c-b399-2c7bcc133893] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 611.613276] env[63202]: DEBUG nova.network.neutron [-] [instance: c9b7182b-8116-419c-b399-2c7bcc133893] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 611.678988] env[63202]: DEBUG nova.network.neutron [-] [instance: c9b7182b-8116-419c-b399-2c7bcc133893] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 611.698105] env[63202]: DEBUG nova.network.neutron [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 611.778640] env[63202]: DEBUG nova.network.neutron [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 611.879256] env[63202]: DEBUG nova.network.neutron [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.004897] env[63202]: INFO nova.compute.manager [-] [instance: 21d0b283-3bb4-4bda-8e62-c933c6de4927] Took 1.06 seconds to deallocate network for instance. [ 612.011831] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] Acquiring lock "4c84555a-83ac-4896-b40c-cbd7e14eebed" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 612.012075] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] Lock "4c84555a-83ac-4896-b40c-cbd7e14eebed" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 612.111068] env[63202]: DEBUG nova.network.neutron [req-efd3fa99-f0b3-4bde-a9a9-2c16601f2415 req-b81eb92f-d1ae-407f-84bb-7fb5f9c83001 service nova] [instance: c9b7182b-8116-419c-b399-2c7bcc133893] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.183918] env[63202]: DEBUG nova.network.neutron [-] [instance: c9b7182b-8116-419c-b399-2c7bcc133893] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.198781] env[63202]: DEBUG nova.network.neutron [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.302893] env[63202]: DEBUG nova.compute.manager [req-8380e0d5-711d-4c60-b9e1-95e3afcf15ad req-746a78e1-0e6c-4098-877e-554fcb1e7c45 service nova] [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] Received event network-changed-684d1759-1f66-45f5-86c1-6fad8a310af1 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 612.303536] env[63202]: DEBUG nova.compute.manager [req-8380e0d5-711d-4c60-b9e1-95e3afcf15ad req-746a78e1-0e6c-4098-877e-554fcb1e7c45 service nova] [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] Refreshing instance network info cache due to event network-changed-684d1759-1f66-45f5-86c1-6fad8a310af1. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 612.304309] env[63202]: DEBUG oslo_concurrency.lockutils [req-8380e0d5-711d-4c60-b9e1-95e3afcf15ad req-746a78e1-0e6c-4098-877e-554fcb1e7c45 service nova] Acquiring lock "refresh_cache-c538fef4-2245-4bd3-af03-f0105fceb1e3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 612.305237] env[63202]: DEBUG oslo_concurrency.lockutils [req-8380e0d5-711d-4c60-b9e1-95e3afcf15ad req-746a78e1-0e6c-4098-877e-554fcb1e7c45 service nova] Acquired lock "refresh_cache-c538fef4-2245-4bd3-af03-f0105fceb1e3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.305237] env[63202]: DEBUG nova.network.neutron [req-8380e0d5-711d-4c60-b9e1-95e3afcf15ad req-746a78e1-0e6c-4098-877e-554fcb1e7c45 service nova] [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] Refreshing network info cache for port 684d1759-1f66-45f5-86c1-6fad8a310af1 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 612.382205] env[63202]: DEBUG oslo_concurrency.lockutils [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] Releasing lock "refresh_cache-6e102794-152e-492e-b063-e1ebaffa82b6" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 612.383363] env[63202]: DEBUG nova.compute.manager [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 612.383363] env[63202]: DEBUG nova.compute.manager [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 612.383460] env[63202]: DEBUG nova.network.neutron [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 612.428825] env[63202]: DEBUG nova.network.neutron [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.516950] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3a242ef3-a939-47fb-8baa-b578de6860c3 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 612.522290] env[63202]: DEBUG nova.network.neutron [req-efd3fa99-f0b3-4bde-a9a9-2c16601f2415 req-b81eb92f-d1ae-407f-84bb-7fb5f9c83001 service nova] [instance: c9b7182b-8116-419c-b399-2c7bcc133893] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.567054] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99dea38c-b1b9-4af4-943f-73d4c002af48 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.580622] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7578c89-591e-4814-8e81-bc23ab1ba248 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.614122] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5b545f7-c36a-4947-8619-e7bcf6234359 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.621873] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06786410-3b4f-484c-aa9c-f5aa28af1c91 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.636833] env[63202]: DEBUG nova.compute.provider_tree [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 612.688118] env[63202]: INFO nova.compute.manager [-] [instance: c9b7182b-8116-419c-b399-2c7bcc133893] Took 1.07 seconds to deallocate network for instance. [ 612.692043] env[63202]: DEBUG nova.compute.claims [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: c9b7182b-8116-419c-b399-2c7bcc133893] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 612.692236] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 612.699329] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Releasing lock "refresh_cache-4f0aba5d-1722-4ddb-8331-a10c25e29b6f" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 612.699721] env[63202]: DEBUG nova.compute.manager [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 612.699949] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 612.700199] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b91222ae-ee06-4bfa-9f80-837c1dc48c50 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.709287] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29a5bc07-b3c6-4add-b192-a043f0cb0032 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.736790] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4f0aba5d-1722-4ddb-8331-a10c25e29b6f could not be found. [ 612.736790] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 612.736790] env[63202]: INFO nova.compute.manager [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 612.736790] env[63202]: DEBUG oslo.service.loopingcall [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 612.736790] env[63202]: DEBUG nova.compute.manager [-] [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 612.736790] env[63202]: DEBUG nova.network.neutron [-] [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 612.781841] env[63202]: DEBUG nova.network.neutron [-] [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.859495] env[63202]: DEBUG nova.network.neutron [req-8380e0d5-711d-4c60-b9e1-95e3afcf15ad req-746a78e1-0e6c-4098-877e-554fcb1e7c45 service nova] [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.938260] env[63202]: DEBUG nova.network.neutron [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.939265] env[63202]: DEBUG oslo_concurrency.lockutils [None req-0e915654-0639-4b23-ab19-a30d36ca6829 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Acquiring lock "0ded82aa-c68d-4d11-9e4d-90ed9a19b708" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 612.939477] env[63202]: DEBUG oslo_concurrency.lockutils [None req-0e915654-0639-4b23-ab19-a30d36ca6829 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Lock "0ded82aa-c68d-4d11-9e4d-90ed9a19b708" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 612.939662] env[63202]: DEBUG oslo_concurrency.lockutils [None req-0e915654-0639-4b23-ab19-a30d36ca6829 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Acquiring lock "0ded82aa-c68d-4d11-9e4d-90ed9a19b708-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 612.940217] env[63202]: DEBUG oslo_concurrency.lockutils [None req-0e915654-0639-4b23-ab19-a30d36ca6829 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Lock "0ded82aa-c68d-4d11-9e4d-90ed9a19b708-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 612.940217] env[63202]: DEBUG oslo_concurrency.lockutils [None req-0e915654-0639-4b23-ab19-a30d36ca6829 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Lock "0ded82aa-c68d-4d11-9e4d-90ed9a19b708-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 612.944447] env[63202]: INFO nova.compute.manager [None req-0e915654-0639-4b23-ab19-a30d36ca6829 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Terminating instance [ 612.947072] env[63202]: DEBUG oslo_concurrency.lockutils [None req-0e915654-0639-4b23-ab19-a30d36ca6829 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Acquiring lock "refresh_cache-0ded82aa-c68d-4d11-9e4d-90ed9a19b708" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 612.947072] env[63202]: DEBUG oslo_concurrency.lockutils [None req-0e915654-0639-4b23-ab19-a30d36ca6829 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Acquired lock "refresh_cache-0ded82aa-c68d-4d11-9e4d-90ed9a19b708" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.947072] env[63202]: DEBUG nova.network.neutron [None req-0e915654-0639-4b23-ab19-a30d36ca6829 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 613.031678] env[63202]: DEBUG oslo_concurrency.lockutils [req-efd3fa99-f0b3-4bde-a9a9-2c16601f2415 req-b81eb92f-d1ae-407f-84bb-7fb5f9c83001 service nova] Releasing lock "refresh_cache-c9b7182b-8116-419c-b399-2c7bcc133893" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 613.143104] env[63202]: DEBUG nova.scheduler.client.report [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 613.285940] env[63202]: DEBUG nova.network.neutron [-] [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.343659] env[63202]: DEBUG nova.network.neutron [req-8380e0d5-711d-4c60-b9e1-95e3afcf15ad req-746a78e1-0e6c-4098-877e-554fcb1e7c45 service nova] [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.442563] env[63202]: INFO nova.compute.manager [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] [instance: 6e102794-152e-492e-b063-e1ebaffa82b6] Took 1.06 seconds to deallocate network for instance. [ 613.503815] env[63202]: DEBUG nova.network.neutron [None req-0e915654-0639-4b23-ab19-a30d36ca6829 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.629685] env[63202]: DEBUG nova.network.neutron [None req-0e915654-0639-4b23-ab19-a30d36ca6829 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.649227] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.492s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.650437] env[63202]: DEBUG nova.compute.manager [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 613.654693] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 16.562s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.654959] env[63202]: DEBUG nova.objects.instance [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63202) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 613.787880] env[63202]: INFO nova.compute.manager [-] [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] Took 1.05 seconds to deallocate network for instance. [ 613.794755] env[63202]: DEBUG nova.compute.claims [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 613.794755] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 613.848695] env[63202]: DEBUG oslo_concurrency.lockutils [req-8380e0d5-711d-4c60-b9e1-95e3afcf15ad req-746a78e1-0e6c-4098-877e-554fcb1e7c45 service nova] Releasing lock "refresh_cache-c538fef4-2245-4bd3-af03-f0105fceb1e3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 613.848695] env[63202]: DEBUG nova.compute.manager [req-8380e0d5-711d-4c60-b9e1-95e3afcf15ad req-746a78e1-0e6c-4098-877e-554fcb1e7c45 service nova] [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] Received event network-vif-deleted-684d1759-1f66-45f5-86c1-6fad8a310af1 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 613.848695] env[63202]: DEBUG nova.compute.manager [req-8380e0d5-711d-4c60-b9e1-95e3afcf15ad req-746a78e1-0e6c-4098-877e-554fcb1e7c45 service nova] [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] Received event network-changed-b35498a4-9923-43db-8dda-1e3802f5777a {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 613.848997] env[63202]: DEBUG nova.compute.manager [req-8380e0d5-711d-4c60-b9e1-95e3afcf15ad req-746a78e1-0e6c-4098-877e-554fcb1e7c45 service nova] [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] Refreshing instance network info cache due to event network-changed-b35498a4-9923-43db-8dda-1e3802f5777a. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 613.848997] env[63202]: DEBUG oslo_concurrency.lockutils [req-8380e0d5-711d-4c60-b9e1-95e3afcf15ad req-746a78e1-0e6c-4098-877e-554fcb1e7c45 service nova] Acquiring lock "refresh_cache-2f757f9b-537e-4bd8-b08d-1da991903552" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 613.850251] env[63202]: DEBUG oslo_concurrency.lockutils [req-8380e0d5-711d-4c60-b9e1-95e3afcf15ad req-746a78e1-0e6c-4098-877e-554fcb1e7c45 service nova] Acquired lock "refresh_cache-2f757f9b-537e-4bd8-b08d-1da991903552" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 613.850251] env[63202]: DEBUG nova.network.neutron [req-8380e0d5-711d-4c60-b9e1-95e3afcf15ad req-746a78e1-0e6c-4098-877e-554fcb1e7c45 service nova] [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] Refreshing network info cache for port b35498a4-9923-43db-8dda-1e3802f5777a {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 614.092507] env[63202]: ERROR nova.compute.manager [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f47aceff-c181-4583-8852-323723733b26, please check neutron logs for more information. [ 614.092507] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 614.092507] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 614.092507] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 614.092507] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 614.092507] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 614.092507] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 614.092507] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 614.092507] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 614.092507] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 614.092507] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 614.092507] env[63202]: ERROR nova.compute.manager raise self.value [ 614.092507] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 614.092507] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 614.092507] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 614.092507] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 614.093555] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 614.093555] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 614.093555] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f47aceff-c181-4583-8852-323723733b26, please check neutron logs for more information. [ 614.093555] env[63202]: ERROR nova.compute.manager [ 614.093555] env[63202]: Traceback (most recent call last): [ 614.093555] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 614.093555] env[63202]: listener.cb(fileno) [ 614.093555] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 614.093555] env[63202]: result = function(*args, **kwargs) [ 614.093555] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 614.093555] env[63202]: return func(*args, **kwargs) [ 614.093555] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 614.093555] env[63202]: raise e [ 614.093555] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 614.093555] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 614.093555] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 614.093555] env[63202]: created_port_ids = self._update_ports_for_instance( [ 614.093555] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 614.093555] env[63202]: with excutils.save_and_reraise_exception(): [ 614.093555] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 614.093555] env[63202]: self.force_reraise() [ 614.093555] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 614.093555] env[63202]: raise self.value [ 614.093555] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 614.093555] env[63202]: updated_port = self._update_port( [ 614.093555] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 614.093555] env[63202]: _ensure_no_port_binding_failure(port) [ 614.093555] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 614.093555] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 614.094996] env[63202]: nova.exception.PortBindingFailed: Binding failed for port f47aceff-c181-4583-8852-323723733b26, please check neutron logs for more information. [ 614.094996] env[63202]: Removing descriptor: 21 [ 614.097551] env[63202]: ERROR nova.compute.manager [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f47aceff-c181-4583-8852-323723733b26, please check neutron logs for more information. [ 614.097551] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] Traceback (most recent call last): [ 614.097551] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 614.097551] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] yield resources [ 614.097551] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 614.097551] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] self.driver.spawn(context, instance, image_meta, [ 614.097551] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 614.097551] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] self._vmops.spawn(context, instance, image_meta, injected_files, [ 614.097551] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 614.097551] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] vm_ref = self.build_virtual_machine(instance, [ 614.097551] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 614.097965] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] vif_infos = vmwarevif.get_vif_info(self._session, [ 614.097965] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 614.097965] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] for vif in network_info: [ 614.097965] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 614.097965] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] return self._sync_wrapper(fn, *args, **kwargs) [ 614.097965] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 614.097965] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] self.wait() [ 614.097965] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 614.097965] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] self[:] = self._gt.wait() [ 614.097965] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 614.097965] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] return self._exit_event.wait() [ 614.097965] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 614.097965] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] result = hub.switch() [ 614.098373] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 614.098373] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] return self.greenlet.switch() [ 614.098373] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 614.098373] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] result = function(*args, **kwargs) [ 614.098373] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 614.098373] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] return func(*args, **kwargs) [ 614.098373] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 614.098373] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] raise e [ 614.098373] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 614.098373] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] nwinfo = self.network_api.allocate_for_instance( [ 614.098373] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 614.098373] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] created_port_ids = self._update_ports_for_instance( [ 614.098373] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 614.098762] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] with excutils.save_and_reraise_exception(): [ 614.098762] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 614.098762] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] self.force_reraise() [ 614.098762] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 614.098762] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] raise self.value [ 614.098762] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 614.098762] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] updated_port = self._update_port( [ 614.098762] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 614.098762] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] _ensure_no_port_binding_failure(port) [ 614.098762] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 614.098762] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] raise exception.PortBindingFailed(port_id=port['id']) [ 614.098762] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] nova.exception.PortBindingFailed: Binding failed for port f47aceff-c181-4583-8852-323723733b26, please check neutron logs for more information. [ 614.098762] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] [ 614.099902] env[63202]: INFO nova.compute.manager [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] Terminating instance [ 614.100685] env[63202]: DEBUG oslo_concurrency.lockutils [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "refresh_cache-29c289ea-703c-4083-b0c4-bbc8b4a92142" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 614.100877] env[63202]: DEBUG oslo_concurrency.lockutils [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquired lock "refresh_cache-29c289ea-703c-4083-b0c4-bbc8b4a92142" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 614.101025] env[63202]: DEBUG nova.network.neutron [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 614.133201] env[63202]: DEBUG oslo_concurrency.lockutils [None req-0e915654-0639-4b23-ab19-a30d36ca6829 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Releasing lock "refresh_cache-0ded82aa-c68d-4d11-9e4d-90ed9a19b708" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 614.133201] env[63202]: DEBUG nova.compute.manager [None req-0e915654-0639-4b23-ab19-a30d36ca6829 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 614.133201] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-0e915654-0639-4b23-ab19-a30d36ca6829 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 614.134180] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa88aec6-14e6-457f-ad40-7fcc8b6949c7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.145977] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e915654-0639-4b23-ab19-a30d36ca6829 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 614.146247] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-297e933e-5f93-4833-a659-2d53f1df76b2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.152947] env[63202]: DEBUG oslo_vmware.api [None req-0e915654-0639-4b23-ab19-a30d36ca6829 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Waiting for the task: (returnval){ [ 614.152947] env[63202]: value = "task-1385313" [ 614.152947] env[63202]: _type = "Task" [ 614.152947] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 614.165357] env[63202]: DEBUG nova.compute.utils [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 614.170308] env[63202]: DEBUG oslo_vmware.api [None req-0e915654-0639-4b23-ab19-a30d36ca6829 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385313, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.171687] env[63202]: DEBUG nova.compute.manager [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 614.171687] env[63202]: DEBUG nova.network.neutron [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 614.257033] env[63202]: DEBUG nova.policy [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '84347a46aa0e46c688f0dfc8b88f2534', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a734788bc5604560b29865b950555f7e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 614.377217] env[63202]: DEBUG nova.network.neutron [req-8380e0d5-711d-4c60-b9e1-95e3afcf15ad req-746a78e1-0e6c-4098-877e-554fcb1e7c45 service nova] [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 614.421375] env[63202]: DEBUG oslo_concurrency.lockutils [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "45fce1e9-b1bb-48aa-9752-cbd011136abd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.421964] env[63202]: DEBUG oslo_concurrency.lockutils [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "45fce1e9-b1bb-48aa-9752-cbd011136abd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.004s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 614.488959] env[63202]: INFO nova.scheduler.client.report [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] Deleted allocations for instance 6e102794-152e-492e-b063-e1ebaffa82b6 [ 614.528714] env[63202]: DEBUG nova.network.neutron [req-8380e0d5-711d-4c60-b9e1-95e3afcf15ad req-746a78e1-0e6c-4098-877e-554fcb1e7c45 service nova] [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.631608] env[63202]: DEBUG nova.network.neutron [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 614.639118] env[63202]: DEBUG nova.network.neutron [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] Successfully created port: a788b884-97a2-4f98-9e03-07eba16c3495 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 614.666981] env[63202]: DEBUG oslo_vmware.api [None req-0e915654-0639-4b23-ab19-a30d36ca6829 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385313, 'name': PowerOffVM_Task, 'duration_secs': 0.152474} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 614.668597] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e915654-0639-4b23-ab19-a30d36ca6829 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 614.668939] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-0e915654-0639-4b23-ab19-a30d36ca6829 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 614.669250] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bc26ee6f-35f6-475d-82a5-52ee0c549028 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.672150] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6e2e7519-2fb4-4e9e-a7f6-cda15b5d512d tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.017s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 614.676433] env[63202]: DEBUG nova.compute.manager [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 614.677875] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.292s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 614.679493] env[63202]: INFO nova.compute.claims [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 614.706343] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-0e915654-0639-4b23-ab19-a30d36ca6829 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 614.707456] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-0e915654-0639-4b23-ab19-a30d36ca6829 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Deleting contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 614.707737] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e915654-0639-4b23-ab19-a30d36ca6829 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Deleting the datastore file [datastore1] 0ded82aa-c68d-4d11-9e4d-90ed9a19b708 {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 614.708460] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5c989232-0fd1-4788-a354-7b65accf67d9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.715520] env[63202]: DEBUG oslo_vmware.api [None req-0e915654-0639-4b23-ab19-a30d36ca6829 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Waiting for the task: (returnval){ [ 614.715520] env[63202]: value = "task-1385315" [ 614.715520] env[63202]: _type = "Task" [ 614.715520] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 614.729915] env[63202]: DEBUG oslo_vmware.api [None req-0e915654-0639-4b23-ab19-a30d36ca6829 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385315, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.824407] env[63202]: DEBUG nova.network.neutron [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.930939] env[63202]: DEBUG nova.compute.manager [req-ca8a5409-12f9-43f6-a6e5-6d83aad0517a req-65da87c8-b0e1-48f6-92d3-763013db462b service nova] [instance: c9b7182b-8116-419c-b399-2c7bcc133893] Received event network-vif-deleted-de336b0b-00ff-4033-bc80-b4b1d8e3fa67 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 615.002009] env[63202]: DEBUG oslo_concurrency.lockutils [None req-14a9be4d-2c4b-4609-a414-9cb914002688 tempest-AttachInterfacesV270Test-809427474 tempest-AttachInterfacesV270Test-809427474-project-member] Lock "6e102794-152e-492e-b063-e1ebaffa82b6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.256s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 615.032683] env[63202]: DEBUG oslo_concurrency.lockutils [req-8380e0d5-711d-4c60-b9e1-95e3afcf15ad req-746a78e1-0e6c-4098-877e-554fcb1e7c45 service nova] Releasing lock "refresh_cache-2f757f9b-537e-4bd8-b08d-1da991903552" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 615.032683] env[63202]: DEBUG nova.compute.manager [req-8380e0d5-711d-4c60-b9e1-95e3afcf15ad req-746a78e1-0e6c-4098-877e-554fcb1e7c45 service nova] [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] Received event network-vif-deleted-b35498a4-9923-43db-8dda-1e3802f5777a {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 615.032683] env[63202]: DEBUG nova.compute.manager [req-8380e0d5-711d-4c60-b9e1-95e3afcf15ad req-746a78e1-0e6c-4098-877e-554fcb1e7c45 service nova] [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] Received event network-changed-fdd4ea5c-13a8-48f2-b391-b5e0c0e7241a {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 615.032683] env[63202]: DEBUG nova.compute.manager [req-8380e0d5-711d-4c60-b9e1-95e3afcf15ad req-746a78e1-0e6c-4098-877e-554fcb1e7c45 service nova] [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] Refreshing instance network info cache due to event network-changed-fdd4ea5c-13a8-48f2-b391-b5e0c0e7241a. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 615.032683] env[63202]: DEBUG oslo_concurrency.lockutils [req-8380e0d5-711d-4c60-b9e1-95e3afcf15ad req-746a78e1-0e6c-4098-877e-554fcb1e7c45 service nova] Acquiring lock "refresh_cache-4f0aba5d-1722-4ddb-8331-a10c25e29b6f" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 615.032852] env[63202]: DEBUG oslo_concurrency.lockutils [req-8380e0d5-711d-4c60-b9e1-95e3afcf15ad req-746a78e1-0e6c-4098-877e-554fcb1e7c45 service nova] Acquired lock "refresh_cache-4f0aba5d-1722-4ddb-8331-a10c25e29b6f" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 615.032913] env[63202]: DEBUG nova.network.neutron [req-8380e0d5-711d-4c60-b9e1-95e3afcf15ad req-746a78e1-0e6c-4098-877e-554fcb1e7c45 service nova] [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] Refreshing network info cache for port fdd4ea5c-13a8-48f2-b391-b5e0c0e7241a {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 615.229404] env[63202]: DEBUG oslo_vmware.api [None req-0e915654-0639-4b23-ab19-a30d36ca6829 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Task: {'id': task-1385315, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.09261} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 615.229700] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e915654-0639-4b23-ab19-a30d36ca6829 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 615.229901] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-0e915654-0639-4b23-ab19-a30d36ca6829 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Deleted contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 615.230114] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-0e915654-0639-4b23-ab19-a30d36ca6829 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 615.230282] env[63202]: INFO nova.compute.manager [None req-0e915654-0639-4b23-ab19-a30d36ca6829 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Took 1.10 seconds to destroy the instance on the hypervisor. [ 615.230520] env[63202]: DEBUG oslo.service.loopingcall [None req-0e915654-0639-4b23-ab19-a30d36ca6829 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 615.230722] env[63202]: DEBUG nova.compute.manager [-] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 615.230838] env[63202]: DEBUG nova.network.neutron [-] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 615.289252] env[63202]: DEBUG nova.network.neutron [-] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 615.327570] env[63202]: DEBUG oslo_concurrency.lockutils [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Releasing lock "refresh_cache-29c289ea-703c-4083-b0c4-bbc8b4a92142" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 615.328578] env[63202]: DEBUG nova.compute.manager [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 615.328797] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 615.329370] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-797d1d4d-f504-4bc2-b9ca-0620c6d33306 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.338616] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ab7bcae-6bb5-4901-8b51-496590e14b59 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.364037] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 29c289ea-703c-4083-b0c4-bbc8b4a92142 could not be found. [ 615.364037] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 615.364037] env[63202]: INFO nova.compute.manager [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] Took 0.03 seconds to destroy the instance on the hypervisor. [ 615.364037] env[63202]: DEBUG oslo.service.loopingcall [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 615.364037] env[63202]: DEBUG nova.compute.manager [-] [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 615.364037] env[63202]: DEBUG nova.network.neutron [-] [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 615.509840] env[63202]: DEBUG nova.compute.manager [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 615.551290] env[63202]: DEBUG nova.network.neutron [-] [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 615.584627] env[63202]: DEBUG nova.network.neutron [req-8380e0d5-711d-4c60-b9e1-95e3afcf15ad req-746a78e1-0e6c-4098-877e-554fcb1e7c45 service nova] [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 615.620743] env[63202]: DEBUG nova.compute.manager [req-8a183cc9-7414-46d4-a952-55d16cf1c41e req-0bf374a9-f60f-4204-ae5b-6dc148565069 service nova] [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] Received event network-vif-deleted-fdd4ea5c-13a8-48f2-b391-b5e0c0e7241a {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 615.698554] env[63202]: DEBUG nova.compute.manager [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 615.729936] env[63202]: DEBUG nova.virt.hardware [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:22:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='1da2521c-ed86-46ec-bf60-9619a7bb838a',id=34,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1530667019',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 615.730215] env[63202]: DEBUG nova.virt.hardware [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 615.730448] env[63202]: DEBUG nova.virt.hardware [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 615.730790] env[63202]: DEBUG nova.virt.hardware [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 615.730874] env[63202]: DEBUG nova.virt.hardware [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 615.731042] env[63202]: DEBUG nova.virt.hardware [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 615.731267] env[63202]: DEBUG nova.virt.hardware [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 615.731581] env[63202]: DEBUG nova.virt.hardware [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 615.731776] env[63202]: DEBUG nova.virt.hardware [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 615.731918] env[63202]: DEBUG nova.virt.hardware [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 615.732257] env[63202]: DEBUG nova.virt.hardware [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 615.733697] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-954360bc-1dc2-4907-b4e0-c598b156e681 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.746659] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] Acquiring lock "ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.746885] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] Lock "ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 615.748201] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f76c28b7-4fd7-4477-be2b-5bd0d6f61e2c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.794817] env[63202]: DEBUG nova.network.neutron [-] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.835310] env[63202]: DEBUG nova.network.neutron [req-8380e0d5-711d-4c60-b9e1-95e3afcf15ad req-746a78e1-0e6c-4098-877e-554fcb1e7c45 service nova] [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.053702] env[63202]: DEBUG nova.network.neutron [-] [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.061683] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 616.085189] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59f307c0-77e9-47d4-89cd-e83631a4cc7d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.100194] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a08f4673-b70e-4064-bebb-5b413d6cf995 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.136683] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dd0160f-814f-46b4-a63c-71d5704ce3bd {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.145942] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8de2da5e-a8d4-4fdb-b17f-f06e9f5d5e80 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.163984] env[63202]: DEBUG nova.compute.provider_tree [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 616.298119] env[63202]: INFO nova.compute.manager [-] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Took 1.07 seconds to deallocate network for instance. [ 616.343578] env[63202]: DEBUG oslo_concurrency.lockutils [req-8380e0d5-711d-4c60-b9e1-95e3afcf15ad req-746a78e1-0e6c-4098-877e-554fcb1e7c45 service nova] Releasing lock "refresh_cache-4f0aba5d-1722-4ddb-8331-a10c25e29b6f" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 616.421341] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Acquiring lock "8cd0415d-dab9-4554-966c-1ea8d822deff" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 616.421544] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Lock "8cd0415d-dab9-4554-966c-1ea8d822deff" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.557010] env[63202]: INFO nova.compute.manager [-] [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] Took 1.19 seconds to deallocate network for instance. [ 616.560825] env[63202]: DEBUG nova.compute.claims [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 616.560881] env[63202]: DEBUG oslo_concurrency.lockutils [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 616.578163] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Acquiring lock "e9b604fb-1d88-425c-92b7-008ddaec7c1a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 616.578418] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Lock "e9b604fb-1d88-425c-92b7-008ddaec7c1a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.667445] env[63202]: DEBUG nova.scheduler.client.report [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 616.809086] env[63202]: DEBUG oslo_concurrency.lockutils [None req-0e915654-0639-4b23-ab19-a30d36ca6829 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.177167] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.499s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 617.177749] env[63202]: DEBUG nova.compute.manager [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 617.182471] env[63202]: DEBUG oslo_concurrency.lockutils [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.873s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 617.689805] env[63202]: DEBUG nova.compute.utils [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 617.696778] env[63202]: DEBUG nova.compute.manager [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 617.696778] env[63202]: DEBUG nova.network.neutron [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 617.881928] env[63202]: DEBUG nova.compute.manager [req-a3a96c01-beff-4dea-9ddd-e8ad1b952384 req-af264fac-b7a4-44b1-a002-f5d82a26bb96 service nova] [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] Received event network-changed-f47aceff-c181-4583-8852-323723733b26 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 617.882190] env[63202]: DEBUG nova.compute.manager [req-a3a96c01-beff-4dea-9ddd-e8ad1b952384 req-af264fac-b7a4-44b1-a002-f5d82a26bb96 service nova] [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] Refreshing instance network info cache due to event network-changed-f47aceff-c181-4583-8852-323723733b26. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 617.882407] env[63202]: DEBUG oslo_concurrency.lockutils [req-a3a96c01-beff-4dea-9ddd-e8ad1b952384 req-af264fac-b7a4-44b1-a002-f5d82a26bb96 service nova] Acquiring lock "refresh_cache-29c289ea-703c-4083-b0c4-bbc8b4a92142" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 617.882535] env[63202]: DEBUG oslo_concurrency.lockutils [req-a3a96c01-beff-4dea-9ddd-e8ad1b952384 req-af264fac-b7a4-44b1-a002-f5d82a26bb96 service nova] Acquired lock "refresh_cache-29c289ea-703c-4083-b0c4-bbc8b4a92142" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 617.882685] env[63202]: DEBUG nova.network.neutron [req-a3a96c01-beff-4dea-9ddd-e8ad1b952384 req-af264fac-b7a4-44b1-a002-f5d82a26bb96 service nova] [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] Refreshing network info cache for port f47aceff-c181-4583-8852-323723733b26 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 617.885976] env[63202]: DEBUG nova.policy [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7fb53d850b2342ebaa64cffee7a4ea09', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '51e84b6e17aa48a3a0855194872c9923', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 618.070104] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74d0c5b6-9b6f-4b31-9a7d-91b107757fc1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.080786] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bbedb33-8997-4377-9418-d546fd71e4f5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.118938] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12112a3e-8e32-4c38-89a7-42e6b9c13271 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.129836] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f48d03d-f713-4e59-b8ca-81e92b62171b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.146217] env[63202]: DEBUG nova.compute.provider_tree [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 618.199698] env[63202]: DEBUG nova.compute.manager [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 618.452349] env[63202]: DEBUG nova.network.neutron [req-a3a96c01-beff-4dea-9ddd-e8ad1b952384 req-af264fac-b7a4-44b1-a002-f5d82a26bb96 service nova] [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 618.650648] env[63202]: DEBUG nova.scheduler.client.report [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 618.813021] env[63202]: DEBUG nova.network.neutron [req-a3a96c01-beff-4dea-9ddd-e8ad1b952384 req-af264fac-b7a4-44b1-a002-f5d82a26bb96 service nova] [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.818932] env[63202]: DEBUG oslo_concurrency.lockutils [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Acquiring lock "c8d27e19-0f3c-4cc3-826b-e07b3a635e5e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 618.819189] env[63202]: DEBUG oslo_concurrency.lockutils [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Lock "c8d27e19-0f3c-4cc3-826b-e07b3a635e5e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.937786] env[63202]: ERROR nova.compute.manager [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a788b884-97a2-4f98-9e03-07eba16c3495, please check neutron logs for more information. [ 618.937786] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 618.937786] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 618.937786] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 618.937786] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 618.937786] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 618.937786] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 618.937786] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 618.937786] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.937786] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 618.937786] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.937786] env[63202]: ERROR nova.compute.manager raise self.value [ 618.937786] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 618.937786] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 618.937786] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.937786] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 618.938262] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.938262] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 618.938262] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a788b884-97a2-4f98-9e03-07eba16c3495, please check neutron logs for more information. [ 618.938262] env[63202]: ERROR nova.compute.manager [ 618.938262] env[63202]: Traceback (most recent call last): [ 618.938262] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 618.938262] env[63202]: listener.cb(fileno) [ 618.938262] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 618.938262] env[63202]: result = function(*args, **kwargs) [ 618.938262] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 618.938262] env[63202]: return func(*args, **kwargs) [ 618.938262] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 618.938262] env[63202]: raise e [ 618.938262] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 618.938262] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 618.938262] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 618.938262] env[63202]: created_port_ids = self._update_ports_for_instance( [ 618.938262] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 618.938262] env[63202]: with excutils.save_and_reraise_exception(): [ 618.938262] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.938262] env[63202]: self.force_reraise() [ 618.938262] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.938262] env[63202]: raise self.value [ 618.938262] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 618.938262] env[63202]: updated_port = self._update_port( [ 618.938262] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.938262] env[63202]: _ensure_no_port_binding_failure(port) [ 618.938262] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.938262] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 618.939108] env[63202]: nova.exception.PortBindingFailed: Binding failed for port a788b884-97a2-4f98-9e03-07eba16c3495, please check neutron logs for more information. [ 618.939108] env[63202]: Removing descriptor: 21 [ 618.939108] env[63202]: ERROR nova.compute.manager [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a788b884-97a2-4f98-9e03-07eba16c3495, please check neutron logs for more information. [ 618.939108] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] Traceback (most recent call last): [ 618.939108] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 618.939108] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] yield resources [ 618.939108] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 618.939108] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] self.driver.spawn(context, instance, image_meta, [ 618.939108] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 618.939108] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 618.939108] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 618.939108] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] vm_ref = self.build_virtual_machine(instance, [ 618.940507] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 618.940507] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] vif_infos = vmwarevif.get_vif_info(self._session, [ 618.940507] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 618.940507] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] for vif in network_info: [ 618.940507] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 618.940507] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] return self._sync_wrapper(fn, *args, **kwargs) [ 618.940507] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 618.940507] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] self.wait() [ 618.940507] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 618.940507] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] self[:] = self._gt.wait() [ 618.940507] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 618.940507] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] return self._exit_event.wait() [ 618.940507] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 618.940899] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] result = hub.switch() [ 618.940899] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 618.940899] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] return self.greenlet.switch() [ 618.940899] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 618.940899] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] result = function(*args, **kwargs) [ 618.940899] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 618.940899] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] return func(*args, **kwargs) [ 618.940899] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 618.940899] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] raise e [ 618.940899] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 618.940899] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] nwinfo = self.network_api.allocate_for_instance( [ 618.940899] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 618.940899] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] created_port_ids = self._update_ports_for_instance( [ 618.941675] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 618.941675] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] with excutils.save_and_reraise_exception(): [ 618.941675] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.941675] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] self.force_reraise() [ 618.941675] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.941675] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] raise self.value [ 618.941675] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 618.941675] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] updated_port = self._update_port( [ 618.941675] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.941675] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] _ensure_no_port_binding_failure(port) [ 618.941675] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.941675] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] raise exception.PortBindingFailed(port_id=port['id']) [ 618.942047] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] nova.exception.PortBindingFailed: Binding failed for port a788b884-97a2-4f98-9e03-07eba16c3495, please check neutron logs for more information. [ 618.942047] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] [ 618.942047] env[63202]: INFO nova.compute.manager [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] Terminating instance [ 618.942463] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Acquiring lock "refresh_cache-88f60509-35cc-4d70-a97c-76dd2761c89b" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.942686] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Acquired lock "refresh_cache-88f60509-35cc-4d70-a97c-76dd2761c89b" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.942907] env[63202]: DEBUG nova.network.neutron [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 619.005253] env[63202]: DEBUG nova.network.neutron [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] Successfully created port: 9c718089-680e-4614-909d-c9fc4e02666a {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 619.156902] env[63202]: DEBUG oslo_concurrency.lockutils [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.975s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 619.157518] env[63202]: ERROR nova.compute.manager [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 180d9c18-47ea-4305-a5b8-ab5abcdf5a1a, please check neutron logs for more information. [ 619.157518] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] Traceback (most recent call last): [ 619.157518] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 619.157518] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] self.driver.spawn(context, instance, image_meta, [ 619.157518] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 619.157518] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 619.157518] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 619.157518] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] vm_ref = self.build_virtual_machine(instance, [ 619.157518] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 619.157518] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] vif_infos = vmwarevif.get_vif_info(self._session, [ 619.157518] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 619.158046] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] for vif in network_info: [ 619.158046] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 619.158046] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] return self._sync_wrapper(fn, *args, **kwargs) [ 619.158046] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 619.158046] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] self.wait() [ 619.158046] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 619.158046] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] self[:] = self._gt.wait() [ 619.158046] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 619.158046] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] return self._exit_event.wait() [ 619.158046] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 619.158046] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] result = hub.switch() [ 619.158046] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 619.158046] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] return self.greenlet.switch() [ 619.158809] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 619.158809] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] result = function(*args, **kwargs) [ 619.158809] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 619.158809] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] return func(*args, **kwargs) [ 619.158809] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 619.158809] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] raise e [ 619.158809] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 619.158809] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] nwinfo = self.network_api.allocate_for_instance( [ 619.158809] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 619.158809] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] created_port_ids = self._update_ports_for_instance( [ 619.158809] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 619.158809] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] with excutils.save_and_reraise_exception(): [ 619.158809] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 619.159689] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] self.force_reraise() [ 619.159689] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 619.159689] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] raise self.value [ 619.159689] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 619.159689] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] updated_port = self._update_port( [ 619.159689] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 619.159689] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] _ensure_no_port_binding_failure(port) [ 619.159689] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 619.159689] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] raise exception.PortBindingFailed(port_id=port['id']) [ 619.159689] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] nova.exception.PortBindingFailed: Binding failed for port 180d9c18-47ea-4305-a5b8-ab5abcdf5a1a, please check neutron logs for more information. [ 619.159689] env[63202]: ERROR nova.compute.manager [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] [ 619.160236] env[63202]: DEBUG nova.compute.utils [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] Binding failed for port 180d9c18-47ea-4305-a5b8-ab5abcdf5a1a, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 619.160637] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.452s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 619.164515] env[63202]: INFO nova.compute.claims [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 619.166404] env[63202]: DEBUG nova.compute.manager [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] Build of instance 90999eb1-8829-4f99-b3b9-8c70ee636cc5 was re-scheduled: Binding failed for port 180d9c18-47ea-4305-a5b8-ab5abcdf5a1a, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 619.169839] env[63202]: DEBUG nova.compute.manager [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 619.170272] env[63202]: DEBUG oslo_concurrency.lockutils [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] Acquiring lock "refresh_cache-90999eb1-8829-4f99-b3b9-8c70ee636cc5" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 619.170463] env[63202]: DEBUG oslo_concurrency.lockutils [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] Acquired lock "refresh_cache-90999eb1-8829-4f99-b3b9-8c70ee636cc5" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.170630] env[63202]: DEBUG nova.network.neutron [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 619.213964] env[63202]: DEBUG nova.compute.manager [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 619.248787] env[63202]: DEBUG nova.virt.hardware [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 619.249231] env[63202]: DEBUG nova.virt.hardware [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 619.250055] env[63202]: DEBUG nova.virt.hardware [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 619.250055] env[63202]: DEBUG nova.virt.hardware [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 619.250055] env[63202]: DEBUG nova.virt.hardware [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 619.250278] env[63202]: DEBUG nova.virt.hardware [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 619.250609] env[63202]: DEBUG nova.virt.hardware [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 619.250910] env[63202]: DEBUG nova.virt.hardware [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 619.251992] env[63202]: DEBUG nova.virt.hardware [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 619.251992] env[63202]: DEBUG nova.virt.hardware [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 619.251992] env[63202]: DEBUG nova.virt.hardware [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 619.253713] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df9c5ddb-7236-41df-ab27-fb4849d71b1f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.263563] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f907be14-504e-4f7c-acff-5924f063c855 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.319904] env[63202]: DEBUG oslo_concurrency.lockutils [req-a3a96c01-beff-4dea-9ddd-e8ad1b952384 req-af264fac-b7a4-44b1-a002-f5d82a26bb96 service nova] Releasing lock "refresh_cache-29c289ea-703c-4083-b0c4-bbc8b4a92142" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 619.320230] env[63202]: DEBUG nova.compute.manager [req-a3a96c01-beff-4dea-9ddd-e8ad1b952384 req-af264fac-b7a4-44b1-a002-f5d82a26bb96 service nova] [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] Received event network-vif-deleted-f47aceff-c181-4583-8852-323723733b26 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 619.526420] env[63202]: DEBUG nova.network.neutron [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 619.775312] env[63202]: DEBUG nova.network.neutron [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 619.860848] env[63202]: DEBUG nova.network.neutron [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.182349] env[63202]: DEBUG nova.network.neutron [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.364988] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Releasing lock "refresh_cache-88f60509-35cc-4d70-a97c-76dd2761c89b" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 620.365556] env[63202]: DEBUG nova.compute.manager [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 620.365876] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 620.369473] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9edc922b-4e62-470f-94b2-1292886c7991 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.379665] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69bd0b24-6d84-471f-9a27-715a88611a11 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.414437] env[63202]: DEBUG oslo_concurrency.lockutils [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Acquiring lock "ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 620.414767] env[63202]: DEBUG oslo_concurrency.lockutils [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Lock "ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 620.416020] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 88f60509-35cc-4d70-a97c-76dd2761c89b could not be found. [ 620.416391] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 620.416818] env[63202]: INFO nova.compute.manager [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] Took 0.05 seconds to destroy the instance on the hypervisor. [ 620.417191] env[63202]: DEBUG oslo.service.loopingcall [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 620.418370] env[63202]: DEBUG nova.compute.manager [-] [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 620.418480] env[63202]: DEBUG nova.network.neutron [-] [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 620.474429] env[63202]: DEBUG nova.network.neutron [-] [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 620.627073] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f27a60b6-7590-4939-bcf0-ce22954b7baa {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.638924] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65c44fd7-de06-4017-9a23-9d6cd232946e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.675105] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaaad242-17fc-4bbc-9340-f1955d0b3157 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.682520] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-105d8d88-56e5-4396-8cbc-cb757322fd31 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.698018] env[63202]: DEBUG oslo_concurrency.lockutils [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] Releasing lock "refresh_cache-90999eb1-8829-4f99-b3b9-8c70ee636cc5" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 620.698018] env[63202]: DEBUG nova.compute.manager [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 620.698018] env[63202]: DEBUG nova.compute.manager [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 620.698018] env[63202]: DEBUG nova.network.neutron [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 620.699841] env[63202]: DEBUG nova.compute.provider_tree [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 620.782853] env[63202]: DEBUG nova.network.neutron [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 620.891216] env[63202]: DEBUG nova.compute.manager [req-7892e381-9c21-41c2-8a41-261a565ee420 req-1efa967f-9d04-4adf-ad59-40a48e8ff0b6 service nova] [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] Received event network-changed-a788b884-97a2-4f98-9e03-07eba16c3495 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 620.891394] env[63202]: DEBUG nova.compute.manager [req-7892e381-9c21-41c2-8a41-261a565ee420 req-1efa967f-9d04-4adf-ad59-40a48e8ff0b6 service nova] [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] Refreshing instance network info cache due to event network-changed-a788b884-97a2-4f98-9e03-07eba16c3495. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 620.891682] env[63202]: DEBUG oslo_concurrency.lockutils [req-7892e381-9c21-41c2-8a41-261a565ee420 req-1efa967f-9d04-4adf-ad59-40a48e8ff0b6 service nova] Acquiring lock "refresh_cache-88f60509-35cc-4d70-a97c-76dd2761c89b" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 620.891838] env[63202]: DEBUG oslo_concurrency.lockutils [req-7892e381-9c21-41c2-8a41-261a565ee420 req-1efa967f-9d04-4adf-ad59-40a48e8ff0b6 service nova] Acquired lock "refresh_cache-88f60509-35cc-4d70-a97c-76dd2761c89b" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 620.891991] env[63202]: DEBUG nova.network.neutron [req-7892e381-9c21-41c2-8a41-261a565ee420 req-1efa967f-9d04-4adf-ad59-40a48e8ff0b6 service nova] [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] Refreshing network info cache for port a788b884-97a2-4f98-9e03-07eba16c3495 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 620.981446] env[63202]: DEBUG nova.network.neutron [-] [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.203579] env[63202]: DEBUG nova.scheduler.client.report [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 621.286496] env[63202]: DEBUG nova.network.neutron [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.426231] env[63202]: DEBUG nova.network.neutron [req-7892e381-9c21-41c2-8a41-261a565ee420 req-1efa967f-9d04-4adf-ad59-40a48e8ff0b6 service nova] [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 621.483014] env[63202]: INFO nova.compute.manager [-] [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] Took 1.06 seconds to deallocate network for instance. [ 621.488968] env[63202]: DEBUG nova.compute.claims [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 621.489280] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 621.508431] env[63202]: ERROR nova.compute.manager [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9c718089-680e-4614-909d-c9fc4e02666a, please check neutron logs for more information. [ 621.508431] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 621.508431] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 621.508431] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 621.508431] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 621.508431] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 621.508431] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 621.508431] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 621.508431] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 621.508431] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 621.508431] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 621.508431] env[63202]: ERROR nova.compute.manager raise self.value [ 621.508431] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 621.508431] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 621.508431] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 621.508431] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 621.508920] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 621.508920] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 621.508920] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9c718089-680e-4614-909d-c9fc4e02666a, please check neutron logs for more information. [ 621.508920] env[63202]: ERROR nova.compute.manager [ 621.508920] env[63202]: Traceback (most recent call last): [ 621.508920] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 621.508920] env[63202]: listener.cb(fileno) [ 621.508920] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 621.508920] env[63202]: result = function(*args, **kwargs) [ 621.508920] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 621.508920] env[63202]: return func(*args, **kwargs) [ 621.508920] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 621.508920] env[63202]: raise e [ 621.508920] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 621.508920] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 621.508920] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 621.508920] env[63202]: created_port_ids = self._update_ports_for_instance( [ 621.508920] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 621.508920] env[63202]: with excutils.save_and_reraise_exception(): [ 621.508920] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 621.508920] env[63202]: self.force_reraise() [ 621.508920] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 621.508920] env[63202]: raise self.value [ 621.508920] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 621.508920] env[63202]: updated_port = self._update_port( [ 621.508920] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 621.508920] env[63202]: _ensure_no_port_binding_failure(port) [ 621.508920] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 621.508920] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 621.509716] env[63202]: nova.exception.PortBindingFailed: Binding failed for port 9c718089-680e-4614-909d-c9fc4e02666a, please check neutron logs for more information. [ 621.509716] env[63202]: Removing descriptor: 16 [ 621.509716] env[63202]: ERROR nova.compute.manager [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9c718089-680e-4614-909d-c9fc4e02666a, please check neutron logs for more information. [ 621.509716] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] Traceback (most recent call last): [ 621.509716] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 621.509716] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] yield resources [ 621.509716] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 621.509716] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] self.driver.spawn(context, instance, image_meta, [ 621.509716] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 621.509716] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 621.509716] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 621.509716] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] vm_ref = self.build_virtual_machine(instance, [ 621.510063] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 621.510063] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] vif_infos = vmwarevif.get_vif_info(self._session, [ 621.510063] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 621.510063] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] for vif in network_info: [ 621.510063] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 621.510063] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] return self._sync_wrapper(fn, *args, **kwargs) [ 621.510063] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 621.510063] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] self.wait() [ 621.510063] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 621.510063] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] self[:] = self._gt.wait() [ 621.510063] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 621.510063] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] return self._exit_event.wait() [ 621.510063] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 621.511261] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] result = hub.switch() [ 621.511261] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 621.511261] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] return self.greenlet.switch() [ 621.511261] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 621.511261] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] result = function(*args, **kwargs) [ 621.511261] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 621.511261] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] return func(*args, **kwargs) [ 621.511261] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 621.511261] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] raise e [ 621.511261] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 621.511261] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] nwinfo = self.network_api.allocate_for_instance( [ 621.511261] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 621.511261] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] created_port_ids = self._update_ports_for_instance( [ 621.511813] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 621.511813] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] with excutils.save_and_reraise_exception(): [ 621.511813] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 621.511813] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] self.force_reraise() [ 621.511813] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 621.511813] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] raise self.value [ 621.511813] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 621.511813] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] updated_port = self._update_port( [ 621.511813] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 621.511813] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] _ensure_no_port_binding_failure(port) [ 621.511813] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 621.511813] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] raise exception.PortBindingFailed(port_id=port['id']) [ 621.514999] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] nova.exception.PortBindingFailed: Binding failed for port 9c718089-680e-4614-909d-c9fc4e02666a, please check neutron logs for more information. [ 621.514999] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] [ 621.514999] env[63202]: INFO nova.compute.manager [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] Terminating instance [ 621.515414] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] Acquiring lock "refresh_cache-1c4fe9af-7a19-4aaf-8758-8d945b2ee57e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 621.515414] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] Acquired lock "refresh_cache-1c4fe9af-7a19-4aaf-8758-8d945b2ee57e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.518010] env[63202]: DEBUG nova.network.neutron [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 621.539350] env[63202]: DEBUG nova.compute.manager [req-9cb2765e-7309-41d7-a5d9-1b395e07e91a req-1e6e3b09-fc5e-4fa9-a5e9-ecc87c753e77 service nova] [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] Received event network-changed-9c718089-680e-4614-909d-c9fc4e02666a {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 621.539575] env[63202]: DEBUG nova.compute.manager [req-9cb2765e-7309-41d7-a5d9-1b395e07e91a req-1e6e3b09-fc5e-4fa9-a5e9-ecc87c753e77 service nova] [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] Refreshing instance network info cache due to event network-changed-9c718089-680e-4614-909d-c9fc4e02666a. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 621.539761] env[63202]: DEBUG oslo_concurrency.lockutils [req-9cb2765e-7309-41d7-a5d9-1b395e07e91a req-1e6e3b09-fc5e-4fa9-a5e9-ecc87c753e77 service nova] Acquiring lock "refresh_cache-1c4fe9af-7a19-4aaf-8758-8d945b2ee57e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 621.614985] env[63202]: DEBUG nova.network.neutron [req-7892e381-9c21-41c2-8a41-261a565ee420 req-1efa967f-9d04-4adf-ad59-40a48e8ff0b6 service nova] [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.709468] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.549s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 621.711017] env[63202]: DEBUG nova.compute.manager [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 621.715987] env[63202]: DEBUG oslo_concurrency.lockutils [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 12.648s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 621.715987] env[63202]: DEBUG nova.objects.instance [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63202) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 621.790509] env[63202]: INFO nova.compute.manager [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] [instance: 90999eb1-8829-4f99-b3b9-8c70ee636cc5] Took 1.09 seconds to deallocate network for instance. [ 622.103725] env[63202]: DEBUG nova.network.neutron [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 622.119632] env[63202]: DEBUG oslo_concurrency.lockutils [req-7892e381-9c21-41c2-8a41-261a565ee420 req-1efa967f-9d04-4adf-ad59-40a48e8ff0b6 service nova] Releasing lock "refresh_cache-88f60509-35cc-4d70-a97c-76dd2761c89b" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 622.120009] env[63202]: DEBUG nova.compute.manager [req-7892e381-9c21-41c2-8a41-261a565ee420 req-1efa967f-9d04-4adf-ad59-40a48e8ff0b6 service nova] [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] Received event network-vif-deleted-a788b884-97a2-4f98-9e03-07eba16c3495 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 622.224025] env[63202]: DEBUG nova.compute.utils [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 622.236987] env[63202]: DEBUG nova.compute.manager [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 622.236987] env[63202]: DEBUG nova.network.neutron [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 622.323525] env[63202]: DEBUG nova.policy [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6d4132b49f7e47719de892f6fdae0a7f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e391b265051f434195be46813032b79e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 622.466316] env[63202]: DEBUG nova.network.neutron [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.599633] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] Acquiring lock "864e1f52-f77e-4596-a5b5-dff3679f911f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.600171] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] Lock "864e1f52-f77e-4596-a5b5-dff3679f911f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.735663] env[63202]: DEBUG nova.compute.manager [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 622.742916] env[63202]: DEBUG oslo_concurrency.lockutils [None req-674793e4-5bfc-4da2-b171-92d6bb4be65e tempest-ServersAdmin275Test-1316318664 tempest-ServersAdmin275Test-1316318664-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.026s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 622.742916] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.955s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.853474] env[63202]: INFO nova.scheduler.client.report [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] Deleted allocations for instance 90999eb1-8829-4f99-b3b9-8c70ee636cc5 [ 622.972755] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] Releasing lock "refresh_cache-1c4fe9af-7a19-4aaf-8758-8d945b2ee57e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 622.972755] env[63202]: DEBUG nova.compute.manager [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 622.972755] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 622.972755] env[63202]: DEBUG oslo_concurrency.lockutils [req-9cb2765e-7309-41d7-a5d9-1b395e07e91a req-1e6e3b09-fc5e-4fa9-a5e9-ecc87c753e77 service nova] Acquired lock "refresh_cache-1c4fe9af-7a19-4aaf-8758-8d945b2ee57e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 622.973125] env[63202]: DEBUG nova.network.neutron [req-9cb2765e-7309-41d7-a5d9-1b395e07e91a req-1e6e3b09-fc5e-4fa9-a5e9-ecc87c753e77 service nova] [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] Refreshing network info cache for port 9c718089-680e-4614-909d-c9fc4e02666a {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 622.975054] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4fc11090-de69-450d-b548-861a5d1e8df4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.987787] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-152d83a8-fdcf-45a1-97d0-cee4500ea326 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.015762] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e could not be found. [ 623.016046] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 623.016228] env[63202]: INFO nova.compute.manager [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 623.017032] env[63202]: DEBUG oslo.service.loopingcall [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 623.017032] env[63202]: DEBUG nova.compute.manager [-] [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 623.017032] env[63202]: DEBUG nova.network.neutron [-] [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 623.045608] env[63202]: DEBUG nova.network.neutron [-] [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 623.052829] env[63202]: DEBUG oslo_concurrency.lockutils [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] Acquiring lock "eff8282a-b3cd-4038-9f48-94cd4c4afc55" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.052829] env[63202]: DEBUG oslo_concurrency.lockutils [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] Lock "eff8282a-b3cd-4038-9f48-94cd4c4afc55" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.279154] env[63202]: DEBUG nova.network.neutron [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] Successfully created port: b018dd36-fcdf-4b74-b182-9bbf4024d22a {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 623.359921] env[63202]: DEBUG oslo_concurrency.lockutils [None req-41f8b9c1-cd29-4553-86ea-b158a4818c68 tempest-ImagesNegativeTestJSON-880519392 tempest-ImagesNegativeTestJSON-880519392-project-member] Lock "90999eb1-8829-4f99-b3b9-8c70ee636cc5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.054s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.553515] env[63202]: DEBUG nova.network.neutron [-] [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.648687] env[63202]: DEBUG nova.network.neutron [req-9cb2765e-7309-41d7-a5d9-1b395e07e91a req-1e6e3b09-fc5e-4fa9-a5e9-ecc87c753e77 service nova] [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 623.677776] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ec73cc4-9a79-40f5-9d10-1d5642e09da4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.692047] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4d19d3f-2e20-4b53-8695-d156eb173672 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.730553] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77f95044-e617-402f-af5b-80cec8d61b8e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.738008] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74c72a82-9f75-4d8a-817a-a31ff934ba59 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.752486] env[63202]: DEBUG nova.compute.provider_tree [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 623.761896] env[63202]: DEBUG nova.compute.manager [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 623.792952] env[63202]: DEBUG nova.virt.hardware [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 623.792952] env[63202]: DEBUG nova.virt.hardware [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 623.792952] env[63202]: DEBUG nova.virt.hardware [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 623.793147] env[63202]: DEBUG nova.virt.hardware [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 623.793147] env[63202]: DEBUG nova.virt.hardware [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 623.793439] env[63202]: DEBUG nova.virt.hardware [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 623.794270] env[63202]: DEBUG nova.virt.hardware [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 623.795429] env[63202]: DEBUG nova.virt.hardware [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 623.795429] env[63202]: DEBUG nova.virt.hardware [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 623.795429] env[63202]: DEBUG nova.virt.hardware [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 623.795429] env[63202]: DEBUG nova.virt.hardware [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 623.797021] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8af6a54b-d80d-4125-9eab-3d78fee362d5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.806198] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeab1e0b-a6d2-4eaa-a1bb-a770d28ba51b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.866740] env[63202]: DEBUG nova.compute.manager [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 623.987745] env[63202]: DEBUG nova.network.neutron [req-9cb2765e-7309-41d7-a5d9-1b395e07e91a req-1e6e3b09-fc5e-4fa9-a5e9-ecc87c753e77 service nova] [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 624.056159] env[63202]: INFO nova.compute.manager [-] [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] Took 1.04 seconds to deallocate network for instance. [ 624.060625] env[63202]: DEBUG nova.compute.claims [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 624.060882] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.255777] env[63202]: DEBUG nova.scheduler.client.report [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 624.403297] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.490669] env[63202]: DEBUG oslo_concurrency.lockutils [req-9cb2765e-7309-41d7-a5d9-1b395e07e91a req-1e6e3b09-fc5e-4fa9-a5e9-ecc87c753e77 service nova] Releasing lock "refresh_cache-1c4fe9af-7a19-4aaf-8758-8d945b2ee57e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 624.606757] env[63202]: DEBUG oslo_concurrency.lockutils [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] Acquiring lock "de5f9555-3959-41cd-b63a-db128a001631" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.607019] env[63202]: DEBUG oslo_concurrency.lockutils [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] Lock "de5f9555-3959-41cd-b63a-db128a001631" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.764969] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.022s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 624.765834] env[63202]: ERROR nova.compute.manager [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b35498a4-9923-43db-8dda-1e3802f5777a, please check neutron logs for more information. [ 624.765834] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] Traceback (most recent call last): [ 624.765834] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 624.765834] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] self.driver.spawn(context, instance, image_meta, [ 624.765834] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 624.765834] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] self._vmops.spawn(context, instance, image_meta, injected_files, [ 624.765834] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 624.765834] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] vm_ref = self.build_virtual_machine(instance, [ 624.765834] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 624.765834] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] vif_infos = vmwarevif.get_vif_info(self._session, [ 624.765834] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 624.766232] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] for vif in network_info: [ 624.766232] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 624.766232] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] return self._sync_wrapper(fn, *args, **kwargs) [ 624.766232] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 624.766232] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] self.wait() [ 624.766232] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 624.766232] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] self[:] = self._gt.wait() [ 624.766232] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 624.766232] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] return self._exit_event.wait() [ 624.766232] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 624.766232] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] result = hub.switch() [ 624.766232] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 624.766232] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] return self.greenlet.switch() [ 624.766644] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 624.766644] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] result = function(*args, **kwargs) [ 624.766644] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 624.766644] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] return func(*args, **kwargs) [ 624.766644] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 624.766644] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] raise e [ 624.766644] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 624.766644] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] nwinfo = self.network_api.allocate_for_instance( [ 624.766644] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 624.766644] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] created_port_ids = self._update_ports_for_instance( [ 624.766644] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 624.766644] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] with excutils.save_and_reraise_exception(): [ 624.766644] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.768033] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] self.force_reraise() [ 624.768033] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.768033] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] raise self.value [ 624.768033] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 624.768033] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] updated_port = self._update_port( [ 624.768033] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.768033] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] _ensure_no_port_binding_failure(port) [ 624.768033] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.768033] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] raise exception.PortBindingFailed(port_id=port['id']) [ 624.768033] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] nova.exception.PortBindingFailed: Binding failed for port b35498a4-9923-43db-8dda-1e3802f5777a, please check neutron logs for more information. [ 624.768033] env[63202]: ERROR nova.compute.manager [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] [ 624.768561] env[63202]: DEBUG nova.compute.utils [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] Binding failed for port b35498a4-9923-43db-8dda-1e3802f5777a, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 624.769174] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.569s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.774188] env[63202]: DEBUG nova.compute.manager [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] Build of instance 2f757f9b-537e-4bd8-b08d-1da991903552 was re-scheduled: Binding failed for port b35498a4-9923-43db-8dda-1e3802f5777a, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 624.774188] env[63202]: DEBUG nova.compute.manager [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 624.774188] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] Acquiring lock "refresh_cache-2f757f9b-537e-4bd8-b08d-1da991903552" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.774188] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] Acquired lock "refresh_cache-2f757f9b-537e-4bd8-b08d-1da991903552" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.774400] env[63202]: DEBUG nova.network.neutron [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 624.996940] env[63202]: DEBUG nova.compute.manager [req-4b6d02e2-4006-446f-8b93-d394ba71defa req-b862e250-ce15-48f7-8b36-9132579ca287 service nova] [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] Received event network-vif-deleted-9c718089-680e-4614-909d-c9fc4e02666a {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 625.247117] env[63202]: DEBUG oslo_concurrency.lockutils [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] Acquiring lock "2f87939c-d615-4d13-8695-971e83ca7843" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 625.247382] env[63202]: DEBUG oslo_concurrency.lockutils [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] Lock "2f87939c-d615-4d13-8695-971e83ca7843" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.307139] env[63202]: DEBUG nova.network.neutron [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 625.436454] env[63202]: DEBUG nova.network.neutron [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.739198] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67db2d23-4ae3-432f-98a5-001e62a082ad {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.748361] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db941906-6096-44b5-a52d-cce5d1cf748a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.788105] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3efbdf84-a75a-4ee9-aa46-b549fdad9764 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.797196] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b515dc9-dae5-4a2d-a840-968c527c8be3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.812822] env[63202]: DEBUG nova.compute.provider_tree [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 625.939189] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] Releasing lock "refresh_cache-2f757f9b-537e-4bd8-b08d-1da991903552" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 625.939430] env[63202]: DEBUG nova.compute.manager [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 625.939607] env[63202]: DEBUG nova.compute.manager [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 625.939776] env[63202]: DEBUG nova.network.neutron [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 625.956899] env[63202]: DEBUG nova.network.neutron [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 626.050907] env[63202]: ERROR nova.compute.manager [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b018dd36-fcdf-4b74-b182-9bbf4024d22a, please check neutron logs for more information. [ 626.050907] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 626.050907] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 626.050907] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 626.050907] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 626.050907] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 626.050907] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 626.050907] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 626.050907] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.050907] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 626.050907] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.050907] env[63202]: ERROR nova.compute.manager raise self.value [ 626.050907] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 626.050907] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 626.050907] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.050907] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 626.052477] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.052477] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 626.052477] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b018dd36-fcdf-4b74-b182-9bbf4024d22a, please check neutron logs for more information. [ 626.052477] env[63202]: ERROR nova.compute.manager [ 626.052477] env[63202]: Traceback (most recent call last): [ 626.052477] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 626.052477] env[63202]: listener.cb(fileno) [ 626.052477] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 626.052477] env[63202]: result = function(*args, **kwargs) [ 626.052477] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 626.052477] env[63202]: return func(*args, **kwargs) [ 626.052477] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 626.052477] env[63202]: raise e [ 626.052477] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 626.052477] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 626.052477] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 626.052477] env[63202]: created_port_ids = self._update_ports_for_instance( [ 626.052477] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 626.052477] env[63202]: with excutils.save_and_reraise_exception(): [ 626.052477] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.052477] env[63202]: self.force_reraise() [ 626.052477] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.052477] env[63202]: raise self.value [ 626.052477] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 626.052477] env[63202]: updated_port = self._update_port( [ 626.052477] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.052477] env[63202]: _ensure_no_port_binding_failure(port) [ 626.052477] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.052477] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 626.053433] env[63202]: nova.exception.PortBindingFailed: Binding failed for port b018dd36-fcdf-4b74-b182-9bbf4024d22a, please check neutron logs for more information. [ 626.053433] env[63202]: Removing descriptor: 16 [ 626.053433] env[63202]: ERROR nova.compute.manager [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b018dd36-fcdf-4b74-b182-9bbf4024d22a, please check neutron logs for more information. [ 626.053433] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] Traceback (most recent call last): [ 626.053433] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 626.053433] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] yield resources [ 626.053433] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 626.053433] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] self.driver.spawn(context, instance, image_meta, [ 626.053433] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 626.053433] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 626.053433] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 626.053433] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] vm_ref = self.build_virtual_machine(instance, [ 626.054068] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 626.054068] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] vif_infos = vmwarevif.get_vif_info(self._session, [ 626.054068] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 626.054068] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] for vif in network_info: [ 626.054068] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 626.054068] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] return self._sync_wrapper(fn, *args, **kwargs) [ 626.054068] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 626.054068] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] self.wait() [ 626.054068] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 626.054068] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] self[:] = self._gt.wait() [ 626.054068] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 626.054068] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] return self._exit_event.wait() [ 626.054068] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 626.054468] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] result = hub.switch() [ 626.054468] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 626.054468] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] return self.greenlet.switch() [ 626.054468] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 626.054468] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] result = function(*args, **kwargs) [ 626.054468] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 626.054468] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] return func(*args, **kwargs) [ 626.054468] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 626.054468] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] raise e [ 626.054468] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 626.054468] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] nwinfo = self.network_api.allocate_for_instance( [ 626.054468] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 626.054468] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] created_port_ids = self._update_ports_for_instance( [ 626.054862] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 626.054862] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] with excutils.save_and_reraise_exception(): [ 626.054862] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.054862] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] self.force_reraise() [ 626.054862] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.054862] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] raise self.value [ 626.054862] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 626.054862] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] updated_port = self._update_port( [ 626.054862] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.054862] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] _ensure_no_port_binding_failure(port) [ 626.054862] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.054862] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] raise exception.PortBindingFailed(port_id=port['id']) [ 626.055307] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] nova.exception.PortBindingFailed: Binding failed for port b018dd36-fcdf-4b74-b182-9bbf4024d22a, please check neutron logs for more information. [ 626.055307] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] [ 626.055307] env[63202]: INFO nova.compute.manager [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] Terminating instance [ 626.055307] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] Acquiring lock "refresh_cache-1afb7b0e-7c9a-4594-8214-37fbecc1b7c5" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 626.055307] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] Acquired lock "refresh_cache-1afb7b0e-7c9a-4594-8214-37fbecc1b7c5" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 626.055307] env[63202]: DEBUG nova.network.neutron [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 626.315385] env[63202]: DEBUG nova.scheduler.client.report [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 626.461387] env[63202]: DEBUG nova.network.neutron [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.586130] env[63202]: DEBUG nova.network.neutron [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 626.804164] env[63202]: DEBUG nova.network.neutron [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.822794] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.053s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 626.824015] env[63202]: ERROR nova.compute.manager [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 684d1759-1f66-45f5-86c1-6fad8a310af1, please check neutron logs for more information. [ 626.824015] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] Traceback (most recent call last): [ 626.824015] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 626.824015] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] self.driver.spawn(context, instance, image_meta, [ 626.824015] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 626.824015] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 626.824015] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 626.824015] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] vm_ref = self.build_virtual_machine(instance, [ 626.824015] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 626.824015] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] vif_infos = vmwarevif.get_vif_info(self._session, [ 626.824015] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 626.824424] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] for vif in network_info: [ 626.824424] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 626.824424] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] return self._sync_wrapper(fn, *args, **kwargs) [ 626.824424] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 626.824424] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] self.wait() [ 626.824424] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 626.824424] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] self[:] = self._gt.wait() [ 626.824424] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 626.824424] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] return self._exit_event.wait() [ 626.824424] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 626.824424] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] result = hub.switch() [ 626.824424] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 626.824424] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] return self.greenlet.switch() [ 626.824793] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 626.824793] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] result = function(*args, **kwargs) [ 626.824793] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 626.824793] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] return func(*args, **kwargs) [ 626.824793] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 626.824793] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] raise e [ 626.824793] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 626.824793] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] nwinfo = self.network_api.allocate_for_instance( [ 626.824793] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 626.824793] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] created_port_ids = self._update_ports_for_instance( [ 626.824793] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 626.824793] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] with excutils.save_and_reraise_exception(): [ 626.824793] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.826420] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] self.force_reraise() [ 626.826420] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.826420] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] raise self.value [ 626.826420] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 626.826420] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] updated_port = self._update_port( [ 626.826420] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.826420] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] _ensure_no_port_binding_failure(port) [ 626.826420] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.826420] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] raise exception.PortBindingFailed(port_id=port['id']) [ 626.826420] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] nova.exception.PortBindingFailed: Binding failed for port 684d1759-1f66-45f5-86c1-6fad8a310af1, please check neutron logs for more information. [ 626.826420] env[63202]: ERROR nova.compute.manager [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] [ 626.826844] env[63202]: DEBUG nova.compute.utils [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] Binding failed for port 684d1759-1f66-45f5-86c1-6fad8a310af1, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 626.827348] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3a242ef3-a939-47fb-8baa-b578de6860c3 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.310s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 626.828261] env[63202]: DEBUG nova.objects.instance [None req-3a242ef3-a939-47fb-8baa-b578de6860c3 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Lazy-loading 'resources' on Instance uuid 21d0b283-3bb4-4bda-8e62-c933c6de4927 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 626.829330] env[63202]: DEBUG nova.compute.manager [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] Build of instance c538fef4-2245-4bd3-af03-f0105fceb1e3 was re-scheduled: Binding failed for port 684d1759-1f66-45f5-86c1-6fad8a310af1, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 626.830556] env[63202]: DEBUG nova.compute.manager [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 626.830556] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Acquiring lock "refresh_cache-c538fef4-2245-4bd3-af03-f0105fceb1e3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 626.830556] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Acquired lock "refresh_cache-c538fef4-2245-4bd3-af03-f0105fceb1e3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 626.830556] env[63202]: DEBUG nova.network.neutron [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 626.966370] env[63202]: INFO nova.compute.manager [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] [instance: 2f757f9b-537e-4bd8-b08d-1da991903552] Took 1.03 seconds to deallocate network for instance. [ 627.305620] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] Releasing lock "refresh_cache-1afb7b0e-7c9a-4594-8214-37fbecc1b7c5" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 627.305620] env[63202]: DEBUG nova.compute.manager [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 627.305620] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 627.305876] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-66e499b5-b02c-4caa-9a8e-62368ad42c40 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.317035] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2a732c0-eb7a-4dfb-be49-02e783742123 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.355445] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5 could not be found. [ 627.355755] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 627.356063] env[63202]: INFO nova.compute.manager [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] Took 0.05 seconds to destroy the instance on the hypervisor. [ 627.356325] env[63202]: DEBUG oslo.service.loopingcall [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 627.356543] env[63202]: DEBUG nova.compute.manager [-] [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 627.356632] env[63202]: DEBUG nova.network.neutron [-] [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 627.369021] env[63202]: DEBUG nova.network.neutron [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 627.375348] env[63202]: DEBUG nova.network.neutron [-] [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 627.565395] env[63202]: DEBUG nova.network.neutron [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.624308] env[63202]: DEBUG nova.compute.manager [req-6f2d5be8-e701-4b8c-8a45-0d9ea1cb08c8 req-39f096c7-ecc8-4a73-a69c-9d9004dcceea service nova] [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] Received event network-changed-b018dd36-fcdf-4b74-b182-9bbf4024d22a {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 627.624514] env[63202]: DEBUG nova.compute.manager [req-6f2d5be8-e701-4b8c-8a45-0d9ea1cb08c8 req-39f096c7-ecc8-4a73-a69c-9d9004dcceea service nova] [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] Refreshing instance network info cache due to event network-changed-b018dd36-fcdf-4b74-b182-9bbf4024d22a. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 627.625187] env[63202]: DEBUG oslo_concurrency.lockutils [req-6f2d5be8-e701-4b8c-8a45-0d9ea1cb08c8 req-39f096c7-ecc8-4a73-a69c-9d9004dcceea service nova] Acquiring lock "refresh_cache-1afb7b0e-7c9a-4594-8214-37fbecc1b7c5" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 627.625187] env[63202]: DEBUG oslo_concurrency.lockutils [req-6f2d5be8-e701-4b8c-8a45-0d9ea1cb08c8 req-39f096c7-ecc8-4a73-a69c-9d9004dcceea service nova] Acquired lock "refresh_cache-1afb7b0e-7c9a-4594-8214-37fbecc1b7c5" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.625187] env[63202]: DEBUG nova.network.neutron [req-6f2d5be8-e701-4b8c-8a45-0d9ea1cb08c8 req-39f096c7-ecc8-4a73-a69c-9d9004dcceea service nova] [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] Refreshing network info cache for port b018dd36-fcdf-4b74-b182-9bbf4024d22a {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 627.825673] env[63202]: DEBUG oslo_concurrency.lockutils [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Acquiring lock "8498aa14-fc1a-42e4-a5e6-e22b239a8f7e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 627.825979] env[63202]: DEBUG oslo_concurrency.lockutils [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Lock "8498aa14-fc1a-42e4-a5e6-e22b239a8f7e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 627.879786] env[63202]: DEBUG nova.network.neutron [-] [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.923860] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd9e6120-96d7-42fb-b93c-903c6eb007c7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.935981] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-637c4911-74e9-4227-8a89-901e54d5bbcc {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.971618] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32b50516-0298-4b1d-b93c-e3fbd809faf0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.985129] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7144588-3d0d-4aa5-8049-0e0b637f46d8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.002439] env[63202]: DEBUG nova.compute.provider_tree [None req-3a242ef3-a939-47fb-8baa-b578de6860c3 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 628.011152] env[63202]: INFO nova.scheduler.client.report [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] Deleted allocations for instance 2f757f9b-537e-4bd8-b08d-1da991903552 [ 628.069217] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Releasing lock "refresh_cache-c538fef4-2245-4bd3-af03-f0105fceb1e3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 628.069579] env[63202]: DEBUG nova.compute.manager [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 628.069791] env[63202]: DEBUG nova.compute.manager [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 628.069962] env[63202]: DEBUG nova.network.neutron [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 628.109862] env[63202]: DEBUG nova.network.neutron [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 628.185624] env[63202]: DEBUG nova.network.neutron [req-6f2d5be8-e701-4b8c-8a45-0d9ea1cb08c8 req-39f096c7-ecc8-4a73-a69c-9d9004dcceea service nova] [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 628.344331] env[63202]: DEBUG nova.network.neutron [req-6f2d5be8-e701-4b8c-8a45-0d9ea1cb08c8 req-39f096c7-ecc8-4a73-a69c-9d9004dcceea service nova] [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.385970] env[63202]: INFO nova.compute.manager [-] [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] Took 1.03 seconds to deallocate network for instance. [ 628.389715] env[63202]: DEBUG nova.compute.claims [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 628.389894] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 628.507798] env[63202]: DEBUG nova.scheduler.client.report [None req-3a242ef3-a939-47fb-8baa-b578de6860c3 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 628.518863] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d559edf3-cef8-4f04-ba61-c5e48c66e676 tempest-VolumesAssistedSnapshotsTest-1603003897 tempest-VolumesAssistedSnapshotsTest-1603003897-project-member] Lock "2f757f9b-537e-4bd8-b08d-1da991903552" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.687s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 628.616344] env[63202]: DEBUG nova.network.neutron [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.848300] env[63202]: DEBUG oslo_concurrency.lockutils [req-6f2d5be8-e701-4b8c-8a45-0d9ea1cb08c8 req-39f096c7-ecc8-4a73-a69c-9d9004dcceea service nova] Releasing lock "refresh_cache-1afb7b0e-7c9a-4594-8214-37fbecc1b7c5" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 628.848300] env[63202]: DEBUG nova.compute.manager [req-6f2d5be8-e701-4b8c-8a45-0d9ea1cb08c8 req-39f096c7-ecc8-4a73-a69c-9d9004dcceea service nova] [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] Received event network-vif-deleted-b018dd36-fcdf-4b74-b182-9bbf4024d22a {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 629.013432] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3a242ef3-a939-47fb-8baa-b578de6860c3 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.186s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 629.015859] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.324s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.023604] env[63202]: DEBUG nova.compute.manager [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 629.051035] env[63202]: INFO nova.scheduler.client.report [None req-3a242ef3-a939-47fb-8baa-b578de6860c3 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Deleted allocations for instance 21d0b283-3bb4-4bda-8e62-c933c6de4927 [ 629.116280] env[63202]: INFO nova.compute.manager [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: c538fef4-2245-4bd3-af03-f0105fceb1e3] Took 1.05 seconds to deallocate network for instance. [ 629.566467] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3a242ef3-a939-47fb-8baa-b578de6860c3 tempest-ServerDiagnosticsV248Test-352704655 tempest-ServerDiagnosticsV248Test-352704655-project-member] Lock "21d0b283-3bb4-4bda-8e62-c933c6de4927" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.101s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 629.574896] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.620499] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Acquiring lock "06496aad-b025-48c2-8436-03d43d1ac899" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.620720] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Lock "06496aad-b025-48c2-8436-03d43d1ac899" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.978142] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e446b3d3-5632-49d5-8734-a10dadfec74b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.988833] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef95d4d5-d8cb-4015-a0dc-e2a5b1397350 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.024572] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-219476df-d30b-41cc-a901-d2e496aa69aa {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.032341] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4f1db12-2c4b-4042-becb-de7858a30e83 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.045886] env[63202]: DEBUG nova.compute.provider_tree [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 630.191702] env[63202]: INFO nova.scheduler.client.report [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Deleted allocations for instance c538fef4-2245-4bd3-af03-f0105fceb1e3 [ 630.550759] env[63202]: DEBUG nova.scheduler.client.report [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 630.706977] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5352e9d5-0d1d-43c7-84e1-73d2858f379b tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Lock "c538fef4-2245-4bd3-af03-f0105fceb1e3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.751s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 631.060566] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.045s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 631.061218] env[63202]: ERROR nova.compute.manager [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: c9b7182b-8116-419c-b399-2c7bcc133893] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port de336b0b-00ff-4033-bc80-b4b1d8e3fa67, please check neutron logs for more information. [ 631.061218] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] Traceback (most recent call last): [ 631.061218] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 631.061218] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] self.driver.spawn(context, instance, image_meta, [ 631.061218] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 631.061218] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] self._vmops.spawn(context, instance, image_meta, injected_files, [ 631.061218] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 631.061218] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] vm_ref = self.build_virtual_machine(instance, [ 631.061218] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 631.061218] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] vif_infos = vmwarevif.get_vif_info(self._session, [ 631.061218] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 631.061703] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] for vif in network_info: [ 631.061703] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 631.061703] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] return self._sync_wrapper(fn, *args, **kwargs) [ 631.061703] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 631.061703] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] self.wait() [ 631.061703] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 631.061703] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] self[:] = self._gt.wait() [ 631.061703] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 631.061703] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] return self._exit_event.wait() [ 631.061703] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 631.061703] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] result = hub.switch() [ 631.061703] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 631.061703] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] return self.greenlet.switch() [ 631.062368] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 631.062368] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] result = function(*args, **kwargs) [ 631.062368] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 631.062368] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] return func(*args, **kwargs) [ 631.062368] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 631.062368] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] raise e [ 631.062368] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 631.062368] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] nwinfo = self.network_api.allocate_for_instance( [ 631.062368] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 631.062368] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] created_port_ids = self._update_ports_for_instance( [ 631.062368] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 631.062368] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] with excutils.save_and_reraise_exception(): [ 631.062368] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.062814] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] self.force_reraise() [ 631.062814] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.062814] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] raise self.value [ 631.062814] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 631.062814] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] updated_port = self._update_port( [ 631.062814] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.062814] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] _ensure_no_port_binding_failure(port) [ 631.062814] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.062814] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] raise exception.PortBindingFailed(port_id=port['id']) [ 631.062814] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] nova.exception.PortBindingFailed: Binding failed for port de336b0b-00ff-4033-bc80-b4b1d8e3fa67, please check neutron logs for more information. [ 631.062814] env[63202]: ERROR nova.compute.manager [instance: c9b7182b-8116-419c-b399-2c7bcc133893] [ 631.063317] env[63202]: DEBUG nova.compute.utils [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: c9b7182b-8116-419c-b399-2c7bcc133893] Binding failed for port de336b0b-00ff-4033-bc80-b4b1d8e3fa67, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 631.063399] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.269s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.066352] env[63202]: DEBUG nova.compute.manager [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: c9b7182b-8116-419c-b399-2c7bcc133893] Build of instance c9b7182b-8116-419c-b399-2c7bcc133893 was re-scheduled: Binding failed for port de336b0b-00ff-4033-bc80-b4b1d8e3fa67, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 631.066785] env[63202]: DEBUG nova.compute.manager [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: c9b7182b-8116-419c-b399-2c7bcc133893] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 631.067010] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Acquiring lock "refresh_cache-c9b7182b-8116-419c-b399-2c7bcc133893" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 631.067166] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Acquired lock "refresh_cache-c9b7182b-8116-419c-b399-2c7bcc133893" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.067320] env[63202]: DEBUG nova.network.neutron [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: c9b7182b-8116-419c-b399-2c7bcc133893] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 631.218195] env[63202]: DEBUG nova.compute.manager [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 631.420711] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Acquiring lock "8885156e-78f1-45ff-87c9-829f3ab89dca" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.420962] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Lock "8885156e-78f1-45ff-87c9-829f3ab89dca" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.615244] env[63202]: DEBUG nova.network.neutron [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: c9b7182b-8116-419c-b399-2c7bcc133893] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 631.747471] env[63202]: DEBUG nova.network.neutron [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: c9b7182b-8116-419c-b399-2c7bcc133893] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.756020] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.021700] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f95f23e-fa21-4388-93fe-9a5295d2a37a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.030031] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aee2cdb5-dbcc-4827-9ef0-7ec73420a04d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.078596] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d60e98b-98d5-44c0-9334-5e21046289bb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.089544] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a43c4e88-b75d-427d-85f6-7aabb04e8c5f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.107877] env[63202]: DEBUG nova.compute.provider_tree [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 632.252922] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Releasing lock "refresh_cache-c9b7182b-8116-419c-b399-2c7bcc133893" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 632.253164] env[63202]: DEBUG nova.compute.manager [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 632.253952] env[63202]: DEBUG nova.compute.manager [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: c9b7182b-8116-419c-b399-2c7bcc133893] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 632.253952] env[63202]: DEBUG nova.network.neutron [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: c9b7182b-8116-419c-b399-2c7bcc133893] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 632.279991] env[63202]: DEBUG nova.network.neutron [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: c9b7182b-8116-419c-b399-2c7bcc133893] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 632.614628] env[63202]: DEBUG nova.scheduler.client.report [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 632.783835] env[63202]: DEBUG nova.network.neutron [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: c9b7182b-8116-419c-b399-2c7bcc133893] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.123925] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.060s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 633.124691] env[63202]: ERROR nova.compute.manager [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fdd4ea5c-13a8-48f2-b391-b5e0c0e7241a, please check neutron logs for more information. [ 633.124691] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] Traceback (most recent call last): [ 633.124691] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 633.124691] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] self.driver.spawn(context, instance, image_meta, [ 633.124691] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 633.124691] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 633.124691] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 633.124691] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] vm_ref = self.build_virtual_machine(instance, [ 633.124691] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 633.124691] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] vif_infos = vmwarevif.get_vif_info(self._session, [ 633.124691] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 633.125692] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] for vif in network_info: [ 633.125692] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 633.125692] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] return self._sync_wrapper(fn, *args, **kwargs) [ 633.125692] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 633.125692] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] self.wait() [ 633.125692] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 633.125692] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] self[:] = self._gt.wait() [ 633.125692] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 633.125692] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] return self._exit_event.wait() [ 633.125692] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 633.125692] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] result = hub.switch() [ 633.125692] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 633.125692] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] return self.greenlet.switch() [ 633.128770] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 633.128770] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] result = function(*args, **kwargs) [ 633.128770] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 633.128770] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] return func(*args, **kwargs) [ 633.128770] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 633.128770] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] raise e [ 633.128770] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 633.128770] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] nwinfo = self.network_api.allocate_for_instance( [ 633.128770] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 633.128770] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] created_port_ids = self._update_ports_for_instance( [ 633.128770] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 633.128770] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] with excutils.save_and_reraise_exception(): [ 633.128770] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 633.129884] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] self.force_reraise() [ 633.129884] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 633.129884] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] raise self.value [ 633.129884] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 633.129884] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] updated_port = self._update_port( [ 633.129884] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 633.129884] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] _ensure_no_port_binding_failure(port) [ 633.129884] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 633.129884] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] raise exception.PortBindingFailed(port_id=port['id']) [ 633.129884] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] nova.exception.PortBindingFailed: Binding failed for port fdd4ea5c-13a8-48f2-b391-b5e0c0e7241a, please check neutron logs for more information. [ 633.129884] env[63202]: ERROR nova.compute.manager [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] [ 633.130559] env[63202]: DEBUG nova.compute.utils [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] Binding failed for port fdd4ea5c-13a8-48f2-b391-b5e0c0e7241a, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 633.130559] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.065s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.131972] env[63202]: INFO nova.compute.claims [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 633.136065] env[63202]: DEBUG nova.compute.manager [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] Build of instance 4f0aba5d-1722-4ddb-8331-a10c25e29b6f was re-scheduled: Binding failed for port fdd4ea5c-13a8-48f2-b391-b5e0c0e7241a, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 633.136533] env[63202]: DEBUG nova.compute.manager [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 633.136759] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Acquiring lock "refresh_cache-4f0aba5d-1722-4ddb-8331-a10c25e29b6f" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 633.136899] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Acquired lock "refresh_cache-4f0aba5d-1722-4ddb-8331-a10c25e29b6f" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.140634] env[63202]: DEBUG nova.network.neutron [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 633.292029] env[63202]: INFO nova.compute.manager [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: c9b7182b-8116-419c-b399-2c7bcc133893] Took 1.04 seconds to deallocate network for instance. [ 633.496749] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] Acquiring lock "a97cff65-e350-480e-9891-3317fe05be47" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.496982] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] Lock "a97cff65-e350-480e-9891-3317fe05be47" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.676540] env[63202]: DEBUG nova.network.neutron [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 633.822557] env[63202]: DEBUG nova.network.neutron [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.328178] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Releasing lock "refresh_cache-4f0aba5d-1722-4ddb-8331-a10c25e29b6f" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 634.328536] env[63202]: DEBUG nova.compute.manager [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 634.330846] env[63202]: DEBUG nova.compute.manager [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 634.330846] env[63202]: DEBUG nova.network.neutron [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 634.334474] env[63202]: INFO nova.scheduler.client.report [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Deleted allocations for instance c9b7182b-8116-419c-b399-2c7bcc133893 [ 634.423613] env[63202]: DEBUG nova.network.neutron [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 634.657683] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48c99585-390f-4849-a889-1cfcafcb3ad5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.669013] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f98e6cc3-0e2a-4489-acbd-89ce5c9612a4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.706277] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab6093a5-5972-4293-adb1-477e677a34d4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.715356] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a663086-02d1-4829-9591-dd81247242b9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.732114] env[63202]: DEBUG nova.compute.provider_tree [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 634.843957] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Lock "c9b7182b-8116-419c-b399-2c7bcc133893" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.744s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 634.921691] env[63202]: DEBUG nova.network.neutron [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.237153] env[63202]: DEBUG nova.scheduler.client.report [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 635.348124] env[63202]: DEBUG nova.compute.manager [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 635.424493] env[63202]: INFO nova.compute.manager [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4f0aba5d-1722-4ddb-8331-a10c25e29b6f] Took 1.10 seconds to deallocate network for instance. [ 635.743576] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.617s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 635.743899] env[63202]: DEBUG nova.compute.manager [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 635.748110] env[63202]: DEBUG oslo_concurrency.lockutils [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.186s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 635.890615] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 636.261981] env[63202]: DEBUG nova.compute.utils [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 636.266414] env[63202]: DEBUG nova.compute.manager [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 636.266414] env[63202]: DEBUG nova.network.neutron [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 636.409055] env[63202]: DEBUG nova.policy [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2d292997e4c34051acc645b4a65dc369', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5e914f87450949b1a39866e8cfa3a0eb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 636.465731] env[63202]: INFO nova.scheduler.client.report [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Deleted allocations for instance 4f0aba5d-1722-4ddb-8331-a10c25e29b6f [ 636.666558] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Acquiring lock "4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 636.666783] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Lock "4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 636.719644] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51e44b3d-6ae2-4e0d-9439-433a3ad3d5f8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.732500] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-add717f3-a0f4-4196-94d6-44041d90f4f9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.766517] env[63202]: DEBUG nova.compute.manager [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 636.769915] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-342c8d7d-b59f-44f3-b995-0dbd945b39ab {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.778947] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4801e341-0cdd-40b2-b365-c86bb71be4f4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.794967] env[63202]: DEBUG nova.compute.provider_tree [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 636.985080] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dbda92df-fff0-42d6-8251-e1803fbd049b tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Lock "4f0aba5d-1722-4ddb-8331-a10c25e29b6f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.829s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 637.299217] env[63202]: DEBUG nova.scheduler.client.report [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 637.356434] env[63202]: DEBUG nova.network.neutron [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] Successfully created port: 740b10d8-5849-4e37-91ad-ad3d7b185b16 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 637.489501] env[63202]: DEBUG nova.compute.manager [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 637.780897] env[63202]: DEBUG nova.compute.manager [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 637.806997] env[63202]: DEBUG nova.virt.hardware [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 637.807257] env[63202]: DEBUG nova.virt.hardware [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 637.807403] env[63202]: DEBUG nova.virt.hardware [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 637.807698] env[63202]: DEBUG nova.virt.hardware [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 637.807933] env[63202]: DEBUG nova.virt.hardware [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 637.807933] env[63202]: DEBUG nova.virt.hardware [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 637.808705] env[63202]: DEBUG nova.virt.hardware [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 637.808705] env[63202]: DEBUG nova.virt.hardware [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 637.808779] env[63202]: DEBUG nova.virt.hardware [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 637.808873] env[63202]: DEBUG nova.virt.hardware [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 637.809058] env[63202]: DEBUG nova.virt.hardware [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 637.810053] env[63202]: DEBUG oslo_concurrency.lockutils [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.063s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 637.810620] env[63202]: ERROR nova.compute.manager [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f47aceff-c181-4583-8852-323723733b26, please check neutron logs for more information. [ 637.810620] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] Traceback (most recent call last): [ 637.810620] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 637.810620] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] self.driver.spawn(context, instance, image_meta, [ 637.810620] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 637.810620] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] self._vmops.spawn(context, instance, image_meta, injected_files, [ 637.810620] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 637.810620] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] vm_ref = self.build_virtual_machine(instance, [ 637.810620] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 637.810620] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] vif_infos = vmwarevif.get_vif_info(self._session, [ 637.810620] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 637.811421] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] for vif in network_info: [ 637.811421] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 637.811421] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] return self._sync_wrapper(fn, *args, **kwargs) [ 637.811421] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 637.811421] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] self.wait() [ 637.811421] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 637.811421] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] self[:] = self._gt.wait() [ 637.811421] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 637.811421] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] return self._exit_event.wait() [ 637.811421] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 637.811421] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] result = hub.switch() [ 637.811421] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 637.811421] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] return self.greenlet.switch() [ 637.812100] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 637.812100] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] result = function(*args, **kwargs) [ 637.812100] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 637.812100] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] return func(*args, **kwargs) [ 637.812100] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 637.812100] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] raise e [ 637.812100] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 637.812100] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] nwinfo = self.network_api.allocate_for_instance( [ 637.812100] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 637.812100] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] created_port_ids = self._update_ports_for_instance( [ 637.812100] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 637.812100] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] with excutils.save_and_reraise_exception(): [ 637.812100] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 637.812635] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] self.force_reraise() [ 637.812635] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 637.812635] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] raise self.value [ 637.812635] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 637.812635] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] updated_port = self._update_port( [ 637.812635] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 637.812635] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] _ensure_no_port_binding_failure(port) [ 637.812635] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 637.812635] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] raise exception.PortBindingFailed(port_id=port['id']) [ 637.812635] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] nova.exception.PortBindingFailed: Binding failed for port f47aceff-c181-4583-8852-323723733b26, please check neutron logs for more information. [ 637.812635] env[63202]: ERROR nova.compute.manager [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] [ 637.812952] env[63202]: DEBUG nova.compute.utils [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] Binding failed for port f47aceff-c181-4583-8852-323723733b26, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 637.813480] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fa13e92-6a30-4fd3-9a3e-6ef922fd493d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.816458] env[63202]: DEBUG nova.compute.manager [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] Build of instance 29c289ea-703c-4083-b0c4-bbc8b4a92142 was re-scheduled: Binding failed for port f47aceff-c181-4583-8852-323723733b26, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 637.817047] env[63202]: DEBUG nova.compute.manager [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 637.817123] env[63202]: DEBUG oslo_concurrency.lockutils [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "refresh_cache-29c289ea-703c-4083-b0c4-bbc8b4a92142" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 637.817255] env[63202]: DEBUG oslo_concurrency.lockutils [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquired lock "refresh_cache-29c289ea-703c-4083-b0c4-bbc8b4a92142" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 637.817569] env[63202]: DEBUG nova.network.neutron [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 637.818611] env[63202]: DEBUG oslo_concurrency.lockutils [None req-0e915654-0639-4b23-ab19-a30d36ca6829 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.010s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 637.819063] env[63202]: DEBUG nova.objects.instance [None req-0e915654-0639-4b23-ab19-a30d36ca6829 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Lazy-loading 'resources' on Instance uuid 0ded82aa-c68d-4d11-9e4d-90ed9a19b708 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 637.829651] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-371dc183-f158-446e-be0b-b6911722873c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.024338] env[63202]: DEBUG oslo_concurrency.lockutils [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 638.366529] env[63202]: DEBUG nova.network.neutron [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 638.608353] env[63202]: DEBUG nova.network.neutron [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.731307] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70987d26-74bc-44c2-afef-00d803c4a98d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.740577] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d53f5778-4412-41f4-b434-569175af0a24 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.780038] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b311535c-112d-40f6-b640-921b172f012f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.785550] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] Acquiring lock "611fead3-3d1b-41e4-9579-7ad1a1b754c3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 638.785829] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] Lock "611fead3-3d1b-41e4-9579-7ad1a1b754c3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 638.792960] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac913555-1e2c-4274-b5ac-c41ad3733d6a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.808901] env[63202]: DEBUG nova.compute.provider_tree [None req-0e915654-0639-4b23-ab19-a30d36ca6829 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 639.113891] env[63202]: DEBUG oslo_concurrency.lockutils [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Releasing lock "refresh_cache-29c289ea-703c-4083-b0c4-bbc8b4a92142" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 639.113891] env[63202]: DEBUG nova.compute.manager [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 639.113891] env[63202]: DEBUG nova.compute.manager [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 639.113891] env[63202]: DEBUG nova.network.neutron [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 639.141109] env[63202]: DEBUG nova.network.neutron [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.312745] env[63202]: DEBUG nova.scheduler.client.report [None req-0e915654-0639-4b23-ab19-a30d36ca6829 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 639.457600] env[63202]: ERROR nova.compute.manager [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 740b10d8-5849-4e37-91ad-ad3d7b185b16, please check neutron logs for more information. [ 639.457600] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 639.457600] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 639.457600] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 639.457600] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 639.457600] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 639.457600] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 639.457600] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 639.457600] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 639.457600] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 639.457600] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 639.457600] env[63202]: ERROR nova.compute.manager raise self.value [ 639.457600] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 639.457600] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 639.457600] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 639.457600] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 639.458254] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 639.458254] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 639.458254] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 740b10d8-5849-4e37-91ad-ad3d7b185b16, please check neutron logs for more information. [ 639.458254] env[63202]: ERROR nova.compute.manager [ 639.458254] env[63202]: Traceback (most recent call last): [ 639.458254] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 639.458254] env[63202]: listener.cb(fileno) [ 639.458254] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 639.458254] env[63202]: result = function(*args, **kwargs) [ 639.458254] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 639.458254] env[63202]: return func(*args, **kwargs) [ 639.458254] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 639.458254] env[63202]: raise e [ 639.458254] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 639.458254] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 639.458254] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 639.458254] env[63202]: created_port_ids = self._update_ports_for_instance( [ 639.458254] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 639.458254] env[63202]: with excutils.save_and_reraise_exception(): [ 639.458254] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 639.458254] env[63202]: self.force_reraise() [ 639.458254] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 639.458254] env[63202]: raise self.value [ 639.458254] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 639.458254] env[63202]: updated_port = self._update_port( [ 639.458254] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 639.458254] env[63202]: _ensure_no_port_binding_failure(port) [ 639.458254] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 639.458254] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 639.459364] env[63202]: nova.exception.PortBindingFailed: Binding failed for port 740b10d8-5849-4e37-91ad-ad3d7b185b16, please check neutron logs for more information. [ 639.459364] env[63202]: Removing descriptor: 16 [ 639.459536] env[63202]: DEBUG nova.compute.manager [req-d58fa8e0-3af1-400c-9c60-d9889aae6f46 req-700d347a-db3a-49f2-97e9-6a0d6fc4fecc service nova] [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] Received event network-changed-740b10d8-5849-4e37-91ad-ad3d7b185b16 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 639.459725] env[63202]: DEBUG nova.compute.manager [req-d58fa8e0-3af1-400c-9c60-d9889aae6f46 req-700d347a-db3a-49f2-97e9-6a0d6fc4fecc service nova] [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] Refreshing instance network info cache due to event network-changed-740b10d8-5849-4e37-91ad-ad3d7b185b16. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 639.459927] env[63202]: DEBUG oslo_concurrency.lockutils [req-d58fa8e0-3af1-400c-9c60-d9889aae6f46 req-700d347a-db3a-49f2-97e9-6a0d6fc4fecc service nova] Acquiring lock "refresh_cache-13e91ad9-7fe6-4bbe-8a67-d63a48bfac20" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 639.460080] env[63202]: DEBUG oslo_concurrency.lockutils [req-d58fa8e0-3af1-400c-9c60-d9889aae6f46 req-700d347a-db3a-49f2-97e9-6a0d6fc4fecc service nova] Acquired lock "refresh_cache-13e91ad9-7fe6-4bbe-8a67-d63a48bfac20" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.460236] env[63202]: DEBUG nova.network.neutron [req-d58fa8e0-3af1-400c-9c60-d9889aae6f46 req-700d347a-db3a-49f2-97e9-6a0d6fc4fecc service nova] [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] Refreshing network info cache for port 740b10d8-5849-4e37-91ad-ad3d7b185b16 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 639.463992] env[63202]: ERROR nova.compute.manager [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 740b10d8-5849-4e37-91ad-ad3d7b185b16, please check neutron logs for more information. [ 639.463992] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] Traceback (most recent call last): [ 639.463992] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 639.463992] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] yield resources [ 639.463992] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 639.463992] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] self.driver.spawn(context, instance, image_meta, [ 639.463992] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 639.463992] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] self._vmops.spawn(context, instance, image_meta, injected_files, [ 639.463992] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 639.463992] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] vm_ref = self.build_virtual_machine(instance, [ 639.463992] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 639.464481] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] vif_infos = vmwarevif.get_vif_info(self._session, [ 639.464481] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 639.464481] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] for vif in network_info: [ 639.464481] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 639.464481] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] return self._sync_wrapper(fn, *args, **kwargs) [ 639.464481] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 639.464481] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] self.wait() [ 639.464481] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 639.464481] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] self[:] = self._gt.wait() [ 639.464481] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 639.464481] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] return self._exit_event.wait() [ 639.464481] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 639.464481] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] result = hub.switch() [ 639.465111] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 639.465111] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] return self.greenlet.switch() [ 639.465111] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 639.465111] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] result = function(*args, **kwargs) [ 639.465111] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 639.465111] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] return func(*args, **kwargs) [ 639.465111] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 639.465111] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] raise e [ 639.465111] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 639.465111] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] nwinfo = self.network_api.allocate_for_instance( [ 639.465111] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 639.465111] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] created_port_ids = self._update_ports_for_instance( [ 639.465111] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 639.465635] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] with excutils.save_and_reraise_exception(): [ 639.465635] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 639.465635] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] self.force_reraise() [ 639.465635] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 639.465635] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] raise self.value [ 639.465635] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 639.465635] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] updated_port = self._update_port( [ 639.465635] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 639.465635] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] _ensure_no_port_binding_failure(port) [ 639.465635] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 639.465635] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] raise exception.PortBindingFailed(port_id=port['id']) [ 639.465635] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] nova.exception.PortBindingFailed: Binding failed for port 740b10d8-5849-4e37-91ad-ad3d7b185b16, please check neutron logs for more information. [ 639.465635] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] [ 639.466098] env[63202]: INFO nova.compute.manager [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] Terminating instance [ 639.468191] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquiring lock "refresh_cache-13e91ad9-7fe6-4bbe-8a67-d63a48bfac20" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 639.648032] env[63202]: DEBUG nova.network.neutron [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.823189] env[63202]: DEBUG oslo_concurrency.lockutils [None req-0e915654-0639-4b23-ab19-a30d36ca6829 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.002s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 639.823189] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.334s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.860941] env[63202]: INFO nova.scheduler.client.report [None req-0e915654-0639-4b23-ab19-a30d36ca6829 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Deleted allocations for instance 0ded82aa-c68d-4d11-9e4d-90ed9a19b708 [ 640.011020] env[63202]: DEBUG nova.network.neutron [req-d58fa8e0-3af1-400c-9c60-d9889aae6f46 req-700d347a-db3a-49f2-97e9-6a0d6fc4fecc service nova] [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 640.154665] env[63202]: INFO nova.compute.manager [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 29c289ea-703c-4083-b0c4-bbc8b4a92142] Took 1.04 seconds to deallocate network for instance. [ 640.380629] env[63202]: DEBUG oslo_concurrency.lockutils [None req-0e915654-0639-4b23-ab19-a30d36ca6829 tempest-ServersAdmin275Test-1019513060 tempest-ServersAdmin275Test-1019513060-project-member] Lock "0ded82aa-c68d-4d11-9e4d-90ed9a19b708" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.440s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 640.399995] env[63202]: DEBUG nova.network.neutron [req-d58fa8e0-3af1-400c-9c60-d9889aae6f46 req-700d347a-db3a-49f2-97e9-6a0d6fc4fecc service nova] [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.738800] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3da17e5f-db53-447c-a3ab-d8b04f26b43a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.747972] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c73dd19-eb6f-4f2e-85aa-3caf70a6a0df {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.779338] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fdaac89-d324-453d-8bdd-9d621057b177 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.787771] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-139555b4-f990-4b0e-b8c1-c1e58e8a2fe9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.802887] env[63202]: DEBUG nova.compute.provider_tree [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 640.904880] env[63202]: DEBUG oslo_concurrency.lockutils [req-d58fa8e0-3af1-400c-9c60-d9889aae6f46 req-700d347a-db3a-49f2-97e9-6a0d6fc4fecc service nova] Releasing lock "refresh_cache-13e91ad9-7fe6-4bbe-8a67-d63a48bfac20" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 640.904880] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquired lock "refresh_cache-13e91ad9-7fe6-4bbe-8a67-d63a48bfac20" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.904880] env[63202]: DEBUG nova.network.neutron [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 641.201286] env[63202]: INFO nova.scheduler.client.report [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Deleted allocations for instance 29c289ea-703c-4083-b0c4-bbc8b4a92142 [ 641.308592] env[63202]: DEBUG nova.scheduler.client.report [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 641.439181] env[63202]: DEBUG nova.network.neutron [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 641.575973] env[63202]: DEBUG nova.network.neutron [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 641.629112] env[63202]: DEBUG nova.compute.manager [req-a726ce5d-6cb7-4740-b2dd-dccaa971daf5 req-af0d905f-faa0-441e-ad08-aab7c072e6ee service nova] [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] Received event network-vif-deleted-740b10d8-5849-4e37-91ad-ad3d7b185b16 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 641.710055] env[63202]: DEBUG oslo_concurrency.lockutils [None req-aa1126c9-d852-41b0-a050-af6240a68602 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "29c289ea-703c-4083-b0c4-bbc8b4a92142" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.873s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 641.820031] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.997s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 641.820652] env[63202]: ERROR nova.compute.manager [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a788b884-97a2-4f98-9e03-07eba16c3495, please check neutron logs for more information. [ 641.820652] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] Traceback (most recent call last): [ 641.820652] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 641.820652] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] self.driver.spawn(context, instance, image_meta, [ 641.820652] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 641.820652] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 641.820652] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 641.820652] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] vm_ref = self.build_virtual_machine(instance, [ 641.820652] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 641.820652] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] vif_infos = vmwarevif.get_vif_info(self._session, [ 641.820652] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 641.821075] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] for vif in network_info: [ 641.821075] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 641.821075] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] return self._sync_wrapper(fn, *args, **kwargs) [ 641.821075] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 641.821075] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] self.wait() [ 641.821075] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 641.821075] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] self[:] = self._gt.wait() [ 641.821075] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 641.821075] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] return self._exit_event.wait() [ 641.821075] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 641.821075] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] result = hub.switch() [ 641.821075] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 641.821075] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] return self.greenlet.switch() [ 641.821505] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 641.821505] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] result = function(*args, **kwargs) [ 641.821505] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 641.821505] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] return func(*args, **kwargs) [ 641.821505] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 641.821505] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] raise e [ 641.821505] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 641.821505] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] nwinfo = self.network_api.allocate_for_instance( [ 641.821505] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 641.821505] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] created_port_ids = self._update_ports_for_instance( [ 641.821505] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 641.821505] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] with excutils.save_and_reraise_exception(): [ 641.821505] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 641.821941] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] self.force_reraise() [ 641.821941] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 641.821941] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] raise self.value [ 641.821941] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 641.821941] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] updated_port = self._update_port( [ 641.821941] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 641.821941] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] _ensure_no_port_binding_failure(port) [ 641.821941] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 641.821941] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] raise exception.PortBindingFailed(port_id=port['id']) [ 641.821941] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] nova.exception.PortBindingFailed: Binding failed for port a788b884-97a2-4f98-9e03-07eba16c3495, please check neutron logs for more information. [ 641.821941] env[63202]: ERROR nova.compute.manager [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] [ 641.822323] env[63202]: DEBUG nova.compute.utils [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] Binding failed for port a788b884-97a2-4f98-9e03-07eba16c3495, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 641.825195] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.762s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 641.826277] env[63202]: DEBUG nova.compute.manager [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] Build of instance 88f60509-35cc-4d70-a97c-76dd2761c89b was re-scheduled: Binding failed for port a788b884-97a2-4f98-9e03-07eba16c3495, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 641.826900] env[63202]: DEBUG nova.compute.manager [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 641.827141] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Acquiring lock "refresh_cache-88f60509-35cc-4d70-a97c-76dd2761c89b" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 641.827228] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Acquired lock "refresh_cache-88f60509-35cc-4d70-a97c-76dd2761c89b" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.827497] env[63202]: DEBUG nova.network.neutron [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 641.986113] env[63202]: DEBUG oslo_concurrency.lockutils [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Acquiring lock "4cd7f7b3-d947-4745-8fd7-940076865e3b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 641.986599] env[63202]: DEBUG oslo_concurrency.lockutils [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Lock "4cd7f7b3-d947-4745-8fd7-940076865e3b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.024169] env[63202]: DEBUG oslo_concurrency.lockutils [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Acquiring lock "ba2d047d-d30f-4ba0-bcfb-787c5a3ae516" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 642.024169] env[63202]: DEBUG oslo_concurrency.lockutils [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Lock "ba2d047d-d30f-4ba0-bcfb-787c5a3ae516" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.128930] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Releasing lock "refresh_cache-13e91ad9-7fe6-4bbe-8a67-d63a48bfac20" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 642.128930] env[63202]: DEBUG nova.compute.manager [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 642.128930] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 642.128930] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3f670943-ae95-4089-8964-7ace10eaca3b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.128930] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58f5cb67-c0a4-4e02-a826-e423aafe384f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.131938] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20 could not be found. [ 642.132120] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 642.132355] env[63202]: INFO nova.compute.manager [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] Took 0.05 seconds to destroy the instance on the hypervisor. [ 642.133221] env[63202]: DEBUG oslo.service.loopingcall [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 642.133221] env[63202]: DEBUG nova.compute.manager [-] [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 642.133331] env[63202]: DEBUG nova.network.neutron [-] [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 642.168976] env[63202]: DEBUG nova.network.neutron [-] [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 642.215611] env[63202]: DEBUG nova.compute.manager [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 642.379867] env[63202]: DEBUG nova.network.neutron [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 642.473330] env[63202]: DEBUG nova.network.neutron [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.676080] env[63202]: DEBUG nova.network.neutron [-] [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.736708] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 642.796094] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4f2e2aa-d0c4-4c60-87c0-b196d0279cc4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.804679] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b5f4e15-3a5c-47d7-a011-b63cbd63e2c9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.848504] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de206e64-66c0-41d2-803d-57d8a0bee920 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.858096] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fa07401-426a-498d-89a0-cccb1f1cf51e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.873042] env[63202]: DEBUG nova.compute.provider_tree [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 642.945112] env[63202]: DEBUG oslo_concurrency.lockutils [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Acquiring lock "1ac8e024-65d1-4250-a946-858aaff852c8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 642.945530] env[63202]: DEBUG oslo_concurrency.lockutils [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Lock "1ac8e024-65d1-4250-a946-858aaff852c8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.978823] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Releasing lock "refresh_cache-88f60509-35cc-4d70-a97c-76dd2761c89b" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 642.979506] env[63202]: DEBUG nova.compute.manager [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 642.979506] env[63202]: DEBUG nova.compute.manager [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 642.979710] env[63202]: DEBUG nova.network.neutron [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 643.000336] env[63202]: DEBUG nova.network.neutron [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 643.181023] env[63202]: INFO nova.compute.manager [-] [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] Took 1.05 seconds to deallocate network for instance. [ 643.182506] env[63202]: DEBUG nova.compute.claims [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 643.182687] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 643.376336] env[63202]: DEBUG nova.scheduler.client.report [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 643.502599] env[63202]: DEBUG nova.network.neutron [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.576801] env[63202]: DEBUG oslo_concurrency.lockutils [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Acquiring lock "3080314c-938e-4c27-bffd-547bdc7e6e38" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 643.576801] env[63202]: DEBUG oslo_concurrency.lockutils [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Lock "3080314c-938e-4c27-bffd-547bdc7e6e38" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 643.886020] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.060s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 643.886020] env[63202]: ERROR nova.compute.manager [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9c718089-680e-4614-909d-c9fc4e02666a, please check neutron logs for more information. [ 643.886020] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] Traceback (most recent call last): [ 643.886020] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 643.886020] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] self.driver.spawn(context, instance, image_meta, [ 643.886020] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 643.886020] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 643.886020] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 643.886020] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] vm_ref = self.build_virtual_machine(instance, [ 643.888837] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 643.888837] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] vif_infos = vmwarevif.get_vif_info(self._session, [ 643.888837] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 643.888837] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] for vif in network_info: [ 643.888837] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 643.888837] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] return self._sync_wrapper(fn, *args, **kwargs) [ 643.888837] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 643.888837] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] self.wait() [ 643.888837] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 643.888837] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] self[:] = self._gt.wait() [ 643.888837] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 643.888837] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] return self._exit_event.wait() [ 643.888837] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 643.889170] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] result = hub.switch() [ 643.889170] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 643.889170] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] return self.greenlet.switch() [ 643.889170] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 643.889170] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] result = function(*args, **kwargs) [ 643.889170] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 643.889170] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] return func(*args, **kwargs) [ 643.889170] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 643.889170] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] raise e [ 643.889170] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 643.889170] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] nwinfo = self.network_api.allocate_for_instance( [ 643.889170] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 643.889170] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] created_port_ids = self._update_ports_for_instance( [ 643.889481] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 643.889481] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] with excutils.save_and_reraise_exception(): [ 643.889481] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 643.889481] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] self.force_reraise() [ 643.889481] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 643.889481] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] raise self.value [ 643.889481] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 643.889481] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] updated_port = self._update_port( [ 643.889481] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 643.889481] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] _ensure_no_port_binding_failure(port) [ 643.889481] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 643.889481] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] raise exception.PortBindingFailed(port_id=port['id']) [ 643.889837] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] nova.exception.PortBindingFailed: Binding failed for port 9c718089-680e-4614-909d-c9fc4e02666a, please check neutron logs for more information. [ 643.889837] env[63202]: ERROR nova.compute.manager [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] [ 643.889837] env[63202]: DEBUG nova.compute.utils [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] Binding failed for port 9c718089-680e-4614-909d-c9fc4e02666a, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 643.889837] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.485s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 643.890343] env[63202]: INFO nova.compute.claims [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 643.893319] env[63202]: DEBUG nova.compute.manager [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] Build of instance 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e was re-scheduled: Binding failed for port 9c718089-680e-4614-909d-c9fc4e02666a, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 643.893467] env[63202]: DEBUG nova.compute.manager [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 643.893743] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] Acquiring lock "refresh_cache-1c4fe9af-7a19-4aaf-8758-8d945b2ee57e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 643.893823] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] Acquired lock "refresh_cache-1c4fe9af-7a19-4aaf-8758-8d945b2ee57e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 643.894103] env[63202]: DEBUG nova.network.neutron [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 644.008478] env[63202]: INFO nova.compute.manager [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] [instance: 88f60509-35cc-4d70-a97c-76dd2761c89b] Took 1.03 seconds to deallocate network for instance. [ 644.230527] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] Acquiring lock "7efc454e-337d-43db-9076-bfc5b89eeea4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 644.230907] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] Lock "7efc454e-337d-43db-9076-bfc5b89eeea4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 644.417214] env[63202]: DEBUG nova.network.neutron [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 644.623775] env[63202]: DEBUG nova.network.neutron [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.905621] env[63202]: DEBUG oslo_concurrency.lockutils [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Acquiring lock "fea91da2-186a-44b6-8e20-535e3b123890" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 644.906062] env[63202]: DEBUG oslo_concurrency.lockutils [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Lock "fea91da2-186a-44b6-8e20-535e3b123890" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 645.048269] env[63202]: INFO nova.scheduler.client.report [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Deleted allocations for instance 88f60509-35cc-4d70-a97c-76dd2761c89b [ 645.128998] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] Releasing lock "refresh_cache-1c4fe9af-7a19-4aaf-8758-8d945b2ee57e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 645.130581] env[63202]: DEBUG nova.compute.manager [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 645.133158] env[63202]: DEBUG nova.compute.manager [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 645.133158] env[63202]: DEBUG nova.network.neutron [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 645.165262] env[63202]: DEBUG nova.network.neutron [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 645.435097] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ac3d62d-dc6d-4c91-a561-542ec65802bc {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.444326] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be326d06-18f4-4bb0-9831-268191ae12c3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.105167] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a3a49875-5f09-42a9-80d0-5e6ea8e13650 tempest-MigrationsAdminTest-1373339766 tempest-MigrationsAdminTest-1373339766-project-member] Lock "88f60509-35cc-4d70-a97c-76dd2761c89b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.345s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 646.105477] env[63202]: DEBUG nova.network.neutron [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 646.108918] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1281229d-9571-4ee7-a91d-02a8317c222d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.119885] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a3cf999-39f4-452f-9b73-dee0bc55caaf {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.138745] env[63202]: DEBUG nova.compute.provider_tree [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 646.612408] env[63202]: DEBUG nova.compute.manager [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: 8cd0415d-dab9-4554-966c-1ea8d822deff] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 646.616230] env[63202]: INFO nova.compute.manager [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] [instance: 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e] Took 1.48 seconds to deallocate network for instance. [ 646.646252] env[63202]: DEBUG nova.scheduler.client.report [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 646.782496] env[63202]: DEBUG oslo_concurrency.lockutils [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "8045881e-9bce-46e7-98c6-a7989f61a31e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 646.782681] env[63202]: DEBUG oslo_concurrency.lockutils [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "8045881e-9bce-46e7-98c6-a7989f61a31e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 647.149998] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 647.155042] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.267s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 647.155582] env[63202]: DEBUG nova.compute.manager [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 647.158128] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.768s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 647.668661] env[63202]: DEBUG nova.compute.utils [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 647.673406] env[63202]: INFO nova.scheduler.client.report [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] Deleted allocations for instance 1c4fe9af-7a19-4aaf-8758-8d945b2ee57e [ 647.680366] env[63202]: DEBUG nova.compute.manager [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 647.682430] env[63202]: DEBUG nova.network.neutron [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 647.740971] env[63202]: DEBUG nova.policy [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f68381ee587c436aaf3520ab13f12692', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6088b167b9134db38ec3474e6874cc35', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 648.173949] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96dbf494-ed18-4df7-a5a7-39e0ef0f2d20 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.180892] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d22e574b-375f-44a2-bc33-bc20d96d2860 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409 tempest-FloatingIPsAssociationNegativeTestJSON-1515426409-project-member] Lock "1c4fe9af-7a19-4aaf-8758-8d945b2ee57e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.244s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 648.186158] env[63202]: DEBUG nova.compute.manager [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 648.197867] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59ef2aa1-1b52-44ab-8883-197e4a2be094 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.238822] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2277c56-4dc4-462c-8e5a-aba03f1b07c2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.247791] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6742dda7-7f80-4ce6-8c85-ce36da81c168 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.262014] env[63202]: DEBUG nova.compute.provider_tree [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 648.612415] env[63202]: DEBUG nova.network.neutron [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] Successfully created port: 6da85fe8-7cce-4523-a80d-56a4ee26d1af {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 648.691809] env[63202]: DEBUG nova.compute.manager [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 648.765627] env[63202]: DEBUG nova.scheduler.client.report [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 649.204777] env[63202]: DEBUG nova.compute.manager [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 649.242255] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 649.254184] env[63202]: DEBUG nova.virt.hardware [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 649.254435] env[63202]: DEBUG nova.virt.hardware [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 649.254582] env[63202]: DEBUG nova.virt.hardware [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 649.254780] env[63202]: DEBUG nova.virt.hardware [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 649.254933] env[63202]: DEBUG nova.virt.hardware [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 649.255088] env[63202]: DEBUG nova.virt.hardware [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 649.255305] env[63202]: DEBUG nova.virt.hardware [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 649.255469] env[63202]: DEBUG nova.virt.hardware [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 649.255696] env[63202]: DEBUG nova.virt.hardware [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 649.255904] env[63202]: DEBUG nova.virt.hardware [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 649.256121] env[63202]: DEBUG nova.virt.hardware [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 649.257123] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dca9d7d0-b897-4382-9f75-1a975945a910 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.269053] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e28e5254-77f7-4361-99a9-efebf7d48824 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.290013] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.132s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 649.290840] env[63202]: ERROR nova.compute.manager [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b018dd36-fcdf-4b74-b182-9bbf4024d22a, please check neutron logs for more information. [ 649.290840] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] Traceback (most recent call last): [ 649.290840] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 649.290840] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] self.driver.spawn(context, instance, image_meta, [ 649.290840] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 649.290840] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 649.290840] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 649.290840] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] vm_ref = self.build_virtual_machine(instance, [ 649.290840] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 649.290840] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] vif_infos = vmwarevif.get_vif_info(self._session, [ 649.290840] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 649.291245] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] for vif in network_info: [ 649.291245] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 649.291245] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] return self._sync_wrapper(fn, *args, **kwargs) [ 649.291245] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 649.291245] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] self.wait() [ 649.291245] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 649.291245] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] self[:] = self._gt.wait() [ 649.291245] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 649.291245] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] return self._exit_event.wait() [ 649.291245] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 649.291245] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] result = hub.switch() [ 649.291245] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 649.291245] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] return self.greenlet.switch() [ 649.291518] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 649.291518] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] result = function(*args, **kwargs) [ 649.291518] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 649.291518] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] return func(*args, **kwargs) [ 649.291518] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 649.291518] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] raise e [ 649.291518] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 649.291518] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] nwinfo = self.network_api.allocate_for_instance( [ 649.291518] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 649.291518] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] created_port_ids = self._update_ports_for_instance( [ 649.291518] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 649.291518] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] with excutils.save_and_reraise_exception(): [ 649.291518] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 649.291821] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] self.force_reraise() [ 649.291821] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 649.291821] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] raise self.value [ 649.291821] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 649.291821] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] updated_port = self._update_port( [ 649.291821] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 649.291821] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] _ensure_no_port_binding_failure(port) [ 649.291821] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 649.291821] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] raise exception.PortBindingFailed(port_id=port['id']) [ 649.291821] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] nova.exception.PortBindingFailed: Binding failed for port b018dd36-fcdf-4b74-b182-9bbf4024d22a, please check neutron logs for more information. [ 649.291821] env[63202]: ERROR nova.compute.manager [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] [ 649.292070] env[63202]: DEBUG nova.compute.utils [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] Binding failed for port b018dd36-fcdf-4b74-b182-9bbf4024d22a, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 649.293225] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.718s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 649.294535] env[63202]: INFO nova.compute.claims [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 649.297456] env[63202]: DEBUG nova.compute.manager [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] Build of instance 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5 was re-scheduled: Binding failed for port b018dd36-fcdf-4b74-b182-9bbf4024d22a, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 649.297922] env[63202]: DEBUG nova.compute.manager [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 649.298203] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] Acquiring lock "refresh_cache-1afb7b0e-7c9a-4594-8214-37fbecc1b7c5" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 649.298321] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] Acquired lock "refresh_cache-1afb7b0e-7c9a-4594-8214-37fbecc1b7c5" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 649.298528] env[63202]: DEBUG nova.network.neutron [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 649.649132] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 649.649331] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 649.834554] env[63202]: DEBUG nova.network.neutron [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 650.062024] env[63202]: DEBUG nova.network.neutron [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.156264] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 650.156310] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Starting heal instance info cache {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 650.157140] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Rebuilding the list of instances to heal {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 650.567245] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] Releasing lock "refresh_cache-1afb7b0e-7c9a-4594-8214-37fbecc1b7c5" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 650.567545] env[63202]: DEBUG nova.compute.manager [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 650.567753] env[63202]: DEBUG nova.compute.manager [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 650.568890] env[63202]: DEBUG nova.network.neutron [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 650.665019] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] Skipping network cache update for instance because it is Building. {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 650.665019] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] Skipping network cache update for instance because it is Building. {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 650.665019] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] Skipping network cache update for instance because it is Building. {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 650.665019] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Didn't find any instances for network info cache update. {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 650.665019] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 650.665019] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 650.665249] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 650.665249] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 650.665249] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 650.665249] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 650.665249] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63202) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 650.665249] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager.update_available_resource {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 650.728590] env[63202]: DEBUG nova.network.neutron [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 650.823017] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9ae0b47-e88b-4359-950c-a20baa620e67 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.832076] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ab9a2a9-bfa0-49d6-85c4-57181512cb32 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.863486] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7de63487-ce37-48f6-aa1d-a6c3175f9643 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.875250] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f133d0d-5cd8-49d9-bc96-32f291761146 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.891778] env[63202]: DEBUG nova.compute.provider_tree [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 651.091065] env[63202]: DEBUG oslo_concurrency.lockutils [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] Acquiring lock "e3a6ad78-4f46-42d4-935c-3cf310123530" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.091441] env[63202]: DEBUG oslo_concurrency.lockutils [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] Lock "e3a6ad78-4f46-42d4-935c-3cf310123530" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 651.170024] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.231256] env[63202]: DEBUG nova.network.neutron [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.336526] env[63202]: ERROR nova.compute.manager [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6da85fe8-7cce-4523-a80d-56a4ee26d1af, please check neutron logs for more information. [ 651.336526] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 651.336526] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 651.336526] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 651.336526] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 651.336526] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 651.336526] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 651.336526] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 651.336526] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 651.336526] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 651.336526] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 651.336526] env[63202]: ERROR nova.compute.manager raise self.value [ 651.336526] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 651.336526] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 651.336526] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 651.336526] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 651.336913] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 651.336913] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 651.336913] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6da85fe8-7cce-4523-a80d-56a4ee26d1af, please check neutron logs for more information. [ 651.336913] env[63202]: ERROR nova.compute.manager [ 651.336913] env[63202]: Traceback (most recent call last): [ 651.336913] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 651.336913] env[63202]: listener.cb(fileno) [ 651.336913] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 651.336913] env[63202]: result = function(*args, **kwargs) [ 651.336913] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 651.336913] env[63202]: return func(*args, **kwargs) [ 651.336913] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 651.336913] env[63202]: raise e [ 651.336913] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 651.336913] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 651.336913] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 651.336913] env[63202]: created_port_ids = self._update_ports_for_instance( [ 651.336913] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 651.336913] env[63202]: with excutils.save_and_reraise_exception(): [ 651.336913] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 651.336913] env[63202]: self.force_reraise() [ 651.336913] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 651.336913] env[63202]: raise self.value [ 651.336913] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 651.336913] env[63202]: updated_port = self._update_port( [ 651.336913] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 651.336913] env[63202]: _ensure_no_port_binding_failure(port) [ 651.336913] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 651.336913] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 651.337498] env[63202]: nova.exception.PortBindingFailed: Binding failed for port 6da85fe8-7cce-4523-a80d-56a4ee26d1af, please check neutron logs for more information. [ 651.337498] env[63202]: Removing descriptor: 16 [ 651.337498] env[63202]: ERROR nova.compute.manager [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6da85fe8-7cce-4523-a80d-56a4ee26d1af, please check neutron logs for more information. [ 651.337498] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] Traceback (most recent call last): [ 651.337498] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 651.337498] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] yield resources [ 651.337498] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 651.337498] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] self.driver.spawn(context, instance, image_meta, [ 651.337498] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 651.337498] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] self._vmops.spawn(context, instance, image_meta, injected_files, [ 651.337498] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 651.337498] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] vm_ref = self.build_virtual_machine(instance, [ 651.337838] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 651.337838] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] vif_infos = vmwarevif.get_vif_info(self._session, [ 651.337838] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 651.337838] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] for vif in network_info: [ 651.337838] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 651.337838] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] return self._sync_wrapper(fn, *args, **kwargs) [ 651.337838] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 651.337838] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] self.wait() [ 651.337838] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 651.337838] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] self[:] = self._gt.wait() [ 651.337838] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 651.337838] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] return self._exit_event.wait() [ 651.337838] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 651.338119] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] result = hub.switch() [ 651.338119] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 651.338119] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] return self.greenlet.switch() [ 651.338119] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 651.338119] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] result = function(*args, **kwargs) [ 651.338119] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 651.338119] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] return func(*args, **kwargs) [ 651.338119] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 651.338119] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] raise e [ 651.338119] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 651.338119] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] nwinfo = self.network_api.allocate_for_instance( [ 651.338119] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 651.338119] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] created_port_ids = self._update_ports_for_instance( [ 651.338392] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 651.338392] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] with excutils.save_and_reraise_exception(): [ 651.338392] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 651.338392] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] self.force_reraise() [ 651.338392] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 651.338392] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] raise self.value [ 651.338392] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 651.338392] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] updated_port = self._update_port( [ 651.338392] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 651.338392] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] _ensure_no_port_binding_failure(port) [ 651.338392] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 651.338392] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] raise exception.PortBindingFailed(port_id=port['id']) [ 651.338678] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] nova.exception.PortBindingFailed: Binding failed for port 6da85fe8-7cce-4523-a80d-56a4ee26d1af, please check neutron logs for more information. [ 651.338678] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] [ 651.338678] env[63202]: INFO nova.compute.manager [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] Terminating instance [ 651.340561] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] Acquiring lock "refresh_cache-03e9f5cc-344c-45c8-abff-aae8db411832" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 651.340561] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] Acquired lock "refresh_cache-03e9f5cc-344c-45c8-abff-aae8db411832" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 651.340561] env[63202]: DEBUG nova.network.neutron [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 651.398634] env[63202]: DEBUG nova.scheduler.client.report [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 651.609152] env[63202]: DEBUG nova.compute.manager [req-fc75f032-d081-4aac-9327-3096bc349a6d req-d075041e-14e3-4177-9bf8-54a45e604c56 service nova] [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] Received event network-changed-6da85fe8-7cce-4523-a80d-56a4ee26d1af {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 651.609152] env[63202]: DEBUG nova.compute.manager [req-fc75f032-d081-4aac-9327-3096bc349a6d req-d075041e-14e3-4177-9bf8-54a45e604c56 service nova] [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] Refreshing instance network info cache due to event network-changed-6da85fe8-7cce-4523-a80d-56a4ee26d1af. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 651.610105] env[63202]: DEBUG oslo_concurrency.lockutils [req-fc75f032-d081-4aac-9327-3096bc349a6d req-d075041e-14e3-4177-9bf8-54a45e604c56 service nova] Acquiring lock "refresh_cache-03e9f5cc-344c-45c8-abff-aae8db411832" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 651.734868] env[63202]: INFO nova.compute.manager [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] [instance: 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5] Took 1.17 seconds to deallocate network for instance. [ 651.858859] env[63202]: DEBUG nova.network.neutron [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 651.904175] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.611s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 651.904761] env[63202]: DEBUG nova.compute.manager [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 651.910235] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.153s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 651.911503] env[63202]: INFO nova.compute.claims [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 652.029402] env[63202]: DEBUG nova.network.neutron [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 652.213287] env[63202]: DEBUG oslo_concurrency.lockutils [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Acquiring lock "c9e0ee31-de9e-420d-8dad-380391d6f4e3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 652.213393] env[63202]: DEBUG oslo_concurrency.lockutils [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Lock "c9e0ee31-de9e-420d-8dad-380391d6f4e3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 652.417158] env[63202]: DEBUG nova.compute.utils [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 652.419894] env[63202]: DEBUG nova.compute.manager [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 652.424026] env[63202]: DEBUG nova.network.neutron [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 652.515959] env[63202]: DEBUG nova.policy [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '26a62f0b64fa47bb805fca725f623a57', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e661cefa1ff947fbb61edb5a35cde95c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 652.530390] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] Releasing lock "refresh_cache-03e9f5cc-344c-45c8-abff-aae8db411832" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 652.531084] env[63202]: DEBUG nova.compute.manager [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 652.531084] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 652.531316] env[63202]: DEBUG oslo_concurrency.lockutils [req-fc75f032-d081-4aac-9327-3096bc349a6d req-d075041e-14e3-4177-9bf8-54a45e604c56 service nova] Acquired lock "refresh_cache-03e9f5cc-344c-45c8-abff-aae8db411832" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 652.531478] env[63202]: DEBUG nova.network.neutron [req-fc75f032-d081-4aac-9327-3096bc349a6d req-d075041e-14e3-4177-9bf8-54a45e604c56 service nova] [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] Refreshing network info cache for port 6da85fe8-7cce-4523-a80d-56a4ee26d1af {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 652.532565] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5390fe51-9b28-43ed-a29d-c8c87a8af0da {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.544473] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b83971d4-2f7a-41a1-bdf4-88caf6d2303f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.579646] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 03e9f5cc-344c-45c8-abff-aae8db411832 could not be found. [ 652.579817] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 652.583681] env[63202]: INFO nova.compute.manager [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] Took 0.05 seconds to destroy the instance on the hypervisor. [ 652.583681] env[63202]: DEBUG oslo.service.loopingcall [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 652.583681] env[63202]: DEBUG nova.compute.manager [-] [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 652.583681] env[63202]: DEBUG nova.network.neutron [-] [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 652.596384] env[63202]: DEBUG nova.network.neutron [-] [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 652.778318] env[63202]: INFO nova.scheduler.client.report [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] Deleted allocations for instance 1afb7b0e-7c9a-4594-8214-37fbecc1b7c5 [ 652.926652] env[63202]: DEBUG nova.network.neutron [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] Successfully created port: bde67ad1-7993-4c30-9aa9-13ea5c2f1800 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 652.930913] env[63202]: DEBUG nova.compute.manager [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 652.983233] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Acquiring lock "e24b7994-3786-4709-bcb5-e0b2c9731f2a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 652.983233] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Lock "e24b7994-3786-4709-bcb5-e0b2c9731f2a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 653.062118] env[63202]: DEBUG nova.network.neutron [req-fc75f032-d081-4aac-9327-3096bc349a6d req-d075041e-14e3-4177-9bf8-54a45e604c56 service nova] [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 653.101027] env[63202]: DEBUG nova.network.neutron [-] [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.162392] env[63202]: DEBUG nova.network.neutron [req-fc75f032-d081-4aac-9327-3096bc349a6d req-d075041e-14e3-4177-9bf8-54a45e604c56 service nova] [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.288897] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1a152616-1018-4a26-b217-c857cd5ae93b tempest-AttachInterfacesUnderV243Test-1290048036 tempest-AttachInterfacesUnderV243Test-1290048036-project-member] Lock "1afb7b0e-7c9a-4594-8214-37fbecc1b7c5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 54.011s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 653.455107] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc751665-3534-456a-9ee8-7be132d3aa41 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.464380] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6b6c811-eced-49cc-9501-9fbfb8ce1f48 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.500810] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfe19032-fc59-4747-be67-94c41446945d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.510859] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c214dd4-ea61-4ae7-8632-cc07f83b4c99 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.525197] env[63202]: DEBUG nova.compute.provider_tree [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 653.605059] env[63202]: INFO nova.compute.manager [-] [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] Took 1.02 seconds to deallocate network for instance. [ 653.608590] env[63202]: DEBUG nova.compute.claims [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 653.608933] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 653.665625] env[63202]: DEBUG oslo_concurrency.lockutils [req-fc75f032-d081-4aac-9327-3096bc349a6d req-d075041e-14e3-4177-9bf8-54a45e604c56 service nova] Releasing lock "refresh_cache-03e9f5cc-344c-45c8-abff-aae8db411832" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 653.795509] env[63202]: DEBUG nova.compute.manager [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 653.961965] env[63202]: DEBUG nova.compute.manager [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 653.984328] env[63202]: DEBUG nova.virt.hardware [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:23:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='415383511',id=37,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-1026978613',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 653.984791] env[63202]: DEBUG nova.virt.hardware [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 653.985438] env[63202]: DEBUG nova.virt.hardware [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 653.985809] env[63202]: DEBUG nova.virt.hardware [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 653.986592] env[63202]: DEBUG nova.virt.hardware [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 653.987100] env[63202]: DEBUG nova.virt.hardware [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 653.987565] env[63202]: DEBUG nova.virt.hardware [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 653.987915] env[63202]: DEBUG nova.virt.hardware [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 653.988301] env[63202]: DEBUG nova.virt.hardware [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 653.988623] env[63202]: DEBUG nova.virt.hardware [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 653.989026] env[63202]: DEBUG nova.virt.hardware [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 653.990033] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10d0488a-110e-4b55-a774-3b44ec9ca333 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.000453] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25c82667-6880-4552-bd55-749b5783d37b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.028422] env[63202]: DEBUG nova.scheduler.client.report [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 654.065625] env[63202]: DEBUG nova.compute.manager [req-191c1b72-6c9a-4c80-b03a-d25438499121 req-41b7906e-f141-45f1-85ff-baddec4fad31 service nova] [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] Received event network-vif-deleted-6da85fe8-7cce-4523-a80d-56a4ee26d1af {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 654.323542] env[63202]: DEBUG oslo_concurrency.lockutils [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 654.515150] env[63202]: ERROR nova.compute.manager [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port bde67ad1-7993-4c30-9aa9-13ea5c2f1800, please check neutron logs for more information. [ 654.515150] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 654.515150] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 654.515150] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 654.515150] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 654.515150] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 654.515150] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 654.515150] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 654.515150] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 654.515150] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 654.515150] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 654.515150] env[63202]: ERROR nova.compute.manager raise self.value [ 654.515150] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 654.515150] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 654.515150] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 654.515150] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 654.515617] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 654.515617] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 654.515617] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port bde67ad1-7993-4c30-9aa9-13ea5c2f1800, please check neutron logs for more information. [ 654.515617] env[63202]: ERROR nova.compute.manager [ 654.515617] env[63202]: Traceback (most recent call last): [ 654.515617] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 654.515617] env[63202]: listener.cb(fileno) [ 654.515617] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 654.515617] env[63202]: result = function(*args, **kwargs) [ 654.515617] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 654.515617] env[63202]: return func(*args, **kwargs) [ 654.515617] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 654.515617] env[63202]: raise e [ 654.515617] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 654.515617] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 654.515617] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 654.515617] env[63202]: created_port_ids = self._update_ports_for_instance( [ 654.515617] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 654.515617] env[63202]: with excutils.save_and_reraise_exception(): [ 654.515617] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 654.515617] env[63202]: self.force_reraise() [ 654.515617] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 654.515617] env[63202]: raise self.value [ 654.515617] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 654.515617] env[63202]: updated_port = self._update_port( [ 654.515617] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 654.515617] env[63202]: _ensure_no_port_binding_failure(port) [ 654.515617] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 654.515617] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 654.516366] env[63202]: nova.exception.PortBindingFailed: Binding failed for port bde67ad1-7993-4c30-9aa9-13ea5c2f1800, please check neutron logs for more information. [ 654.516366] env[63202]: Removing descriptor: 16 [ 654.516366] env[63202]: ERROR nova.compute.manager [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port bde67ad1-7993-4c30-9aa9-13ea5c2f1800, please check neutron logs for more information. [ 654.516366] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] Traceback (most recent call last): [ 654.516366] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 654.516366] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] yield resources [ 654.516366] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 654.516366] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] self.driver.spawn(context, instance, image_meta, [ 654.516366] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 654.516366] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 654.516366] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 654.516366] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] vm_ref = self.build_virtual_machine(instance, [ 654.516634] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 654.516634] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] vif_infos = vmwarevif.get_vif_info(self._session, [ 654.516634] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 654.516634] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] for vif in network_info: [ 654.516634] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 654.516634] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] return self._sync_wrapper(fn, *args, **kwargs) [ 654.516634] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 654.516634] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] self.wait() [ 654.516634] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 654.516634] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] self[:] = self._gt.wait() [ 654.516634] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 654.516634] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] return self._exit_event.wait() [ 654.516634] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 654.516908] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] result = hub.switch() [ 654.516908] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 654.516908] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] return self.greenlet.switch() [ 654.516908] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 654.516908] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] result = function(*args, **kwargs) [ 654.516908] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 654.516908] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] return func(*args, **kwargs) [ 654.516908] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 654.516908] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] raise e [ 654.516908] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 654.516908] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] nwinfo = self.network_api.allocate_for_instance( [ 654.516908] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 654.516908] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] created_port_ids = self._update_ports_for_instance( [ 654.517249] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 654.517249] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] with excutils.save_and_reraise_exception(): [ 654.517249] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 654.517249] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] self.force_reraise() [ 654.517249] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 654.517249] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] raise self.value [ 654.517249] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 654.517249] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] updated_port = self._update_port( [ 654.517249] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 654.517249] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] _ensure_no_port_binding_failure(port) [ 654.517249] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 654.517249] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] raise exception.PortBindingFailed(port_id=port['id']) [ 654.517544] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] nova.exception.PortBindingFailed: Binding failed for port bde67ad1-7993-4c30-9aa9-13ea5c2f1800, please check neutron logs for more information. [ 654.517544] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] [ 654.517544] env[63202]: INFO nova.compute.manager [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] Terminating instance [ 654.519417] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] Acquiring lock "refresh_cache-ef69de02-6e43-4489-a35a-8a9096e54fbd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 654.519567] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] Acquired lock "refresh_cache-ef69de02-6e43-4489-a35a-8a9096e54fbd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 654.519730] env[63202]: DEBUG nova.network.neutron [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 654.533134] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.624s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 654.533638] env[63202]: DEBUG nova.compute.manager [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 654.536290] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.646s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 654.537907] env[63202]: INFO nova.compute.claims [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 655.044037] env[63202]: DEBUG nova.compute.utils [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 655.048211] env[63202]: DEBUG nova.compute.manager [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 655.048471] env[63202]: DEBUG nova.network.neutron [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 655.138183] env[63202]: DEBUG nova.policy [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ba9bdc218e2b44b2a338e799a6fb84a0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fc7e68f0c5e3411a81971d90c03adf52', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 655.229268] env[63202]: DEBUG nova.network.neutron [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 655.451452] env[63202]: DEBUG nova.network.neutron [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.557939] env[63202]: DEBUG nova.compute.manager [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 655.594623] env[63202]: DEBUG nova.network.neutron [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] Successfully created port: 3541508a-02bc-425d-9c91-ef407d837b04 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 655.954630] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] Releasing lock "refresh_cache-ef69de02-6e43-4489-a35a-8a9096e54fbd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 655.955299] env[63202]: DEBUG nova.compute.manager [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 655.955698] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 655.956094] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6039e5c1-0cd7-4902-b7b9-86b428223187 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.971027] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3084bb3-9d59-41fd-858d-95e35216b36f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.006277] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ef69de02-6e43-4489-a35a-8a9096e54fbd could not be found. [ 656.006576] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 656.006781] env[63202]: INFO nova.compute.manager [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] Took 0.05 seconds to destroy the instance on the hypervisor. [ 656.007052] env[63202]: DEBUG oslo.service.loopingcall [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 656.007354] env[63202]: DEBUG nova.compute.manager [-] [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 656.007440] env[63202]: DEBUG nova.network.neutron [-] [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 656.042856] env[63202]: DEBUG nova.network.neutron [-] [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 656.059600] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46516a78-8118-45d7-8627-d960340eb596 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.073494] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fbf2069-c879-4447-a7c6-edb9202fc4f7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.109521] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d707af4-8481-4723-860b-82fa08c8e603 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.117945] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7da392b2-a43d-4257-8b63-d6a07c9b296b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.137221] env[63202]: DEBUG nova.compute.provider_tree [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 656.432898] env[63202]: DEBUG nova.compute.manager [req-47ff6804-5dd4-49fb-bd36-28e9a40e2dcc req-70819b8e-27aa-4bc2-a152-0baacf6e81ec service nova] [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] Received event network-changed-bde67ad1-7993-4c30-9aa9-13ea5c2f1800 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 656.433132] env[63202]: DEBUG nova.compute.manager [req-47ff6804-5dd4-49fb-bd36-28e9a40e2dcc req-70819b8e-27aa-4bc2-a152-0baacf6e81ec service nova] [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] Refreshing instance network info cache due to event network-changed-bde67ad1-7993-4c30-9aa9-13ea5c2f1800. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 656.433370] env[63202]: DEBUG oslo_concurrency.lockutils [req-47ff6804-5dd4-49fb-bd36-28e9a40e2dcc req-70819b8e-27aa-4bc2-a152-0baacf6e81ec service nova] Acquiring lock "refresh_cache-ef69de02-6e43-4489-a35a-8a9096e54fbd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 656.433470] env[63202]: DEBUG oslo_concurrency.lockutils [req-47ff6804-5dd4-49fb-bd36-28e9a40e2dcc req-70819b8e-27aa-4bc2-a152-0baacf6e81ec service nova] Acquired lock "refresh_cache-ef69de02-6e43-4489-a35a-8a9096e54fbd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 656.433631] env[63202]: DEBUG nova.network.neutron [req-47ff6804-5dd4-49fb-bd36-28e9a40e2dcc req-70819b8e-27aa-4bc2-a152-0baacf6e81ec service nova] [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] Refreshing network info cache for port bde67ad1-7993-4c30-9aa9-13ea5c2f1800 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 656.545608] env[63202]: DEBUG nova.network.neutron [-] [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.567575] env[63202]: DEBUG nova.compute.manager [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 656.611180] env[63202]: DEBUG nova.virt.hardware [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 656.611180] env[63202]: DEBUG nova.virt.hardware [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 656.611180] env[63202]: DEBUG nova.virt.hardware [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 656.611473] env[63202]: DEBUG nova.virt.hardware [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 656.612176] env[63202]: DEBUG nova.virt.hardware [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 656.612468] env[63202]: DEBUG nova.virt.hardware [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 656.612790] env[63202]: DEBUG nova.virt.hardware [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 656.613146] env[63202]: DEBUG nova.virt.hardware [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 656.613705] env[63202]: DEBUG nova.virt.hardware [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 656.614095] env[63202]: DEBUG nova.virt.hardware [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 656.614390] env[63202]: DEBUG nova.virt.hardware [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 656.615533] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04b7b318-cadd-4fd2-a758-7c3fe9f1a1d5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.639549] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-079973df-28bd-49f2-b2b5-4ce4a5e6049e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.645199] env[63202]: DEBUG nova.scheduler.client.report [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 656.962828] env[63202]: DEBUG nova.network.neutron [req-47ff6804-5dd4-49fb-bd36-28e9a40e2dcc req-70819b8e-27aa-4bc2-a152-0baacf6e81ec service nova] [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 657.048533] env[63202]: INFO nova.compute.manager [-] [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] Took 1.04 seconds to deallocate network for instance. [ 657.051425] env[63202]: DEBUG nova.compute.claims [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 657.052835] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 657.102159] env[63202]: DEBUG nova.network.neutron [req-47ff6804-5dd4-49fb-bd36-28e9a40e2dcc req-70819b8e-27aa-4bc2-a152-0baacf6e81ec service nova] [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.154457] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.616s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 657.154457] env[63202]: DEBUG nova.compute.manager [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 657.156468] env[63202]: DEBUG oslo_concurrency.lockutils [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.133s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 657.161516] env[63202]: INFO nova.compute.claims [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 657.389269] env[63202]: DEBUG nova.compute.manager [req-1e094753-50f0-49db-9d7c-8530cd77bbbf req-131059e4-1d1f-4a77-9ab1-8766093e8ccb service nova] [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] Received event network-changed-3541508a-02bc-425d-9c91-ef407d837b04 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 657.389498] env[63202]: DEBUG nova.compute.manager [req-1e094753-50f0-49db-9d7c-8530cd77bbbf req-131059e4-1d1f-4a77-9ab1-8766093e8ccb service nova] [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] Refreshing instance network info cache due to event network-changed-3541508a-02bc-425d-9c91-ef407d837b04. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 657.389709] env[63202]: DEBUG oslo_concurrency.lockutils [req-1e094753-50f0-49db-9d7c-8530cd77bbbf req-131059e4-1d1f-4a77-9ab1-8766093e8ccb service nova] Acquiring lock "refresh_cache-bc8fc812-1e66-4e14-a39c-1c226e7fe2b9" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 657.389848] env[63202]: DEBUG oslo_concurrency.lockutils [req-1e094753-50f0-49db-9d7c-8530cd77bbbf req-131059e4-1d1f-4a77-9ab1-8766093e8ccb service nova] Acquired lock "refresh_cache-bc8fc812-1e66-4e14-a39c-1c226e7fe2b9" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.390009] env[63202]: DEBUG nova.network.neutron [req-1e094753-50f0-49db-9d7c-8530cd77bbbf req-131059e4-1d1f-4a77-9ab1-8766093e8ccb service nova] [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] Refreshing network info cache for port 3541508a-02bc-425d-9c91-ef407d837b04 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 657.605857] env[63202]: DEBUG oslo_concurrency.lockutils [req-47ff6804-5dd4-49fb-bd36-28e9a40e2dcc req-70819b8e-27aa-4bc2-a152-0baacf6e81ec service nova] Releasing lock "refresh_cache-ef69de02-6e43-4489-a35a-8a9096e54fbd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 657.606394] env[63202]: DEBUG nova.compute.manager [req-47ff6804-5dd4-49fb-bd36-28e9a40e2dcc req-70819b8e-27aa-4bc2-a152-0baacf6e81ec service nova] [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] Received event network-vif-deleted-bde67ad1-7993-4c30-9aa9-13ea5c2f1800 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 657.666630] env[63202]: DEBUG nova.compute.utils [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 657.672038] env[63202]: DEBUG nova.compute.manager [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 657.672271] env[63202]: DEBUG nova.network.neutron [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 657.683448] env[63202]: ERROR nova.compute.manager [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3541508a-02bc-425d-9c91-ef407d837b04, please check neutron logs for more information. [ 657.683448] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 657.683448] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 657.683448] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 657.683448] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 657.683448] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 657.683448] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 657.683448] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 657.683448] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 657.683448] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 657.683448] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 657.683448] env[63202]: ERROR nova.compute.manager raise self.value [ 657.683448] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 657.683448] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 657.683448] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 657.683448] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 657.683895] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 657.683895] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 657.683895] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3541508a-02bc-425d-9c91-ef407d837b04, please check neutron logs for more information. [ 657.683895] env[63202]: ERROR nova.compute.manager [ 657.683895] env[63202]: Traceback (most recent call last): [ 657.683895] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 657.683895] env[63202]: listener.cb(fileno) [ 657.683895] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 657.683895] env[63202]: result = function(*args, **kwargs) [ 657.683895] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 657.683895] env[63202]: return func(*args, **kwargs) [ 657.683895] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 657.683895] env[63202]: raise e [ 657.683895] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 657.683895] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 657.683895] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 657.683895] env[63202]: created_port_ids = self._update_ports_for_instance( [ 657.683895] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 657.683895] env[63202]: with excutils.save_and_reraise_exception(): [ 657.683895] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 657.683895] env[63202]: self.force_reraise() [ 657.683895] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 657.683895] env[63202]: raise self.value [ 657.683895] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 657.683895] env[63202]: updated_port = self._update_port( [ 657.683895] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 657.683895] env[63202]: _ensure_no_port_binding_failure(port) [ 657.683895] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 657.683895] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 657.684626] env[63202]: nova.exception.PortBindingFailed: Binding failed for port 3541508a-02bc-425d-9c91-ef407d837b04, please check neutron logs for more information. [ 657.684626] env[63202]: Removing descriptor: 20 [ 657.684626] env[63202]: ERROR nova.compute.manager [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3541508a-02bc-425d-9c91-ef407d837b04, please check neutron logs for more information. [ 657.684626] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] Traceback (most recent call last): [ 657.684626] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 657.684626] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] yield resources [ 657.684626] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 657.684626] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] self.driver.spawn(context, instance, image_meta, [ 657.684626] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 657.684626] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 657.684626] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 657.684626] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] vm_ref = self.build_virtual_machine(instance, [ 657.684933] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 657.684933] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] vif_infos = vmwarevif.get_vif_info(self._session, [ 657.684933] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 657.684933] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] for vif in network_info: [ 657.684933] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 657.684933] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] return self._sync_wrapper(fn, *args, **kwargs) [ 657.684933] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 657.684933] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] self.wait() [ 657.684933] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 657.684933] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] self[:] = self._gt.wait() [ 657.684933] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 657.684933] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] return self._exit_event.wait() [ 657.684933] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 657.685254] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] result = hub.switch() [ 657.685254] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 657.685254] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] return self.greenlet.switch() [ 657.685254] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 657.685254] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] result = function(*args, **kwargs) [ 657.685254] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 657.685254] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] return func(*args, **kwargs) [ 657.685254] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 657.685254] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] raise e [ 657.685254] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 657.685254] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] nwinfo = self.network_api.allocate_for_instance( [ 657.685254] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 657.685254] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] created_port_ids = self._update_ports_for_instance( [ 657.685522] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 657.685522] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] with excutils.save_and_reraise_exception(): [ 657.685522] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 657.685522] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] self.force_reraise() [ 657.685522] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 657.685522] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] raise self.value [ 657.685522] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 657.685522] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] updated_port = self._update_port( [ 657.685522] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 657.685522] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] _ensure_no_port_binding_failure(port) [ 657.685522] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 657.685522] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] raise exception.PortBindingFailed(port_id=port['id']) [ 657.685821] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] nova.exception.PortBindingFailed: Binding failed for port 3541508a-02bc-425d-9c91-ef407d837b04, please check neutron logs for more information. [ 657.685821] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] [ 657.685821] env[63202]: INFO nova.compute.manager [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] Terminating instance [ 657.690233] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Acquiring lock "refresh_cache-bc8fc812-1e66-4e14-a39c-1c226e7fe2b9" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 657.764757] env[63202]: DEBUG nova.policy [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '747b3ed3dfb04497aa11701e7498e53c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5ed4b087ff67444e819051277d51b9f1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 657.914321] env[63202]: DEBUG nova.network.neutron [req-1e094753-50f0-49db-9d7c-8530cd77bbbf req-131059e4-1d1f-4a77-9ab1-8766093e8ccb service nova] [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 658.011948] env[63202]: DEBUG nova.network.neutron [req-1e094753-50f0-49db-9d7c-8530cd77bbbf req-131059e4-1d1f-4a77-9ab1-8766093e8ccb service nova] [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.171962] env[63202]: DEBUG nova.compute.manager [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 658.452271] env[63202]: DEBUG nova.network.neutron [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] Successfully created port: fba47d74-c242-4267-a6be-f829d6b2dd6b {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 658.517836] env[63202]: DEBUG oslo_concurrency.lockutils [req-1e094753-50f0-49db-9d7c-8530cd77bbbf req-131059e4-1d1f-4a77-9ab1-8766093e8ccb service nova] Releasing lock "refresh_cache-bc8fc812-1e66-4e14-a39c-1c226e7fe2b9" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 658.518279] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Acquired lock "refresh_cache-bc8fc812-1e66-4e14-a39c-1c226e7fe2b9" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 658.518464] env[63202]: DEBUG nova.network.neutron [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 658.631818] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a372a744-f2d5-405a-bb0c-8a3377cb42c3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.639854] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3088af79-8a32-4575-a528-823e6c69a752 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.684472] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26e13b1e-65aa-49d7-8c61-711699334e17 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.694995] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93f7a305-4fbb-4c60-963a-6185a4e00c9a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.712858] env[63202]: DEBUG nova.compute.provider_tree [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 659.045598] env[63202]: DEBUG nova.network.neutron [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 659.150050] env[63202]: DEBUG nova.network.neutron [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.193025] env[63202]: DEBUG nova.compute.manager [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 659.219452] env[63202]: DEBUG nova.scheduler.client.report [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 659.225436] env[63202]: DEBUG nova.virt.hardware [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 659.226072] env[63202]: DEBUG nova.virt.hardware [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 659.226329] env[63202]: DEBUG nova.virt.hardware [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 659.226580] env[63202]: DEBUG nova.virt.hardware [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 659.229026] env[63202]: DEBUG nova.virt.hardware [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 659.229187] env[63202]: DEBUG nova.virt.hardware [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 659.229400] env[63202]: DEBUG nova.virt.hardware [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 659.229554] env[63202]: DEBUG nova.virt.hardware [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 659.229712] env[63202]: DEBUG nova.virt.hardware [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 659.229868] env[63202]: DEBUG nova.virt.hardware [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 659.230044] env[63202]: DEBUG nova.virt.hardware [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 659.231008] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e0e5f79-6e49-4eb5-91d2-b5c79f66a1db {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.242236] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-563145ff-63c2-4843-936d-5aab8256e1e1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.376330] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] Acquiring lock "d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 659.376559] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] Lock "d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 659.652478] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Releasing lock "refresh_cache-bc8fc812-1e66-4e14-a39c-1c226e7fe2b9" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 659.652946] env[63202]: DEBUG nova.compute.manager [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 659.653276] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 659.653565] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-37e15838-a6dc-48f1-b085-a98837a4dd86 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.666319] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2ae30c0-4d71-45cd-b9cc-f5e03d95775a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.695527] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance bc8fc812-1e66-4e14-a39c-1c226e7fe2b9 could not be found. [ 659.695891] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 659.696259] env[63202]: INFO nova.compute.manager [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] Took 0.04 seconds to destroy the instance on the hypervisor. [ 659.696687] env[63202]: DEBUG oslo.service.loopingcall [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 659.697056] env[63202]: DEBUG nova.compute.manager [-] [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 659.697211] env[63202]: DEBUG nova.network.neutron [-] [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 659.723113] env[63202]: DEBUG nova.network.neutron [-] [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 659.729502] env[63202]: DEBUG nova.compute.manager [req-432c4b3b-4ab6-4253-b083-fe593f62dc87 req-d69918cd-73ad-43f7-ab8e-a56193cab9c5 service nova] [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] Received event network-vif-deleted-3541508a-02bc-425d-9c91-ef407d837b04 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 659.734945] env[63202]: DEBUG oslo_concurrency.lockutils [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.578s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 659.735441] env[63202]: DEBUG nova.compute.manager [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 659.738540] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.002s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 659.742151] env[63202]: INFO nova.compute.claims [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 660.106505] env[63202]: DEBUG nova.compute.manager [req-3e696d7b-265d-4b7d-b17f-5acf574b7747 req-346ac9c1-66b7-4f6d-b0e6-97d7b918a4e5 service nova] [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] Received event network-changed-fba47d74-c242-4267-a6be-f829d6b2dd6b {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 660.108110] env[63202]: DEBUG nova.compute.manager [req-3e696d7b-265d-4b7d-b17f-5acf574b7747 req-346ac9c1-66b7-4f6d-b0e6-97d7b918a4e5 service nova] [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] Refreshing instance network info cache due to event network-changed-fba47d74-c242-4267-a6be-f829d6b2dd6b. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 660.108110] env[63202]: DEBUG oslo_concurrency.lockutils [req-3e696d7b-265d-4b7d-b17f-5acf574b7747 req-346ac9c1-66b7-4f6d-b0e6-97d7b918a4e5 service nova] Acquiring lock "refresh_cache-4c84555a-83ac-4896-b40c-cbd7e14eebed" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.108110] env[63202]: DEBUG oslo_concurrency.lockutils [req-3e696d7b-265d-4b7d-b17f-5acf574b7747 req-346ac9c1-66b7-4f6d-b0e6-97d7b918a4e5 service nova] Acquired lock "refresh_cache-4c84555a-83ac-4896-b40c-cbd7e14eebed" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.108110] env[63202]: DEBUG nova.network.neutron [req-3e696d7b-265d-4b7d-b17f-5acf574b7747 req-346ac9c1-66b7-4f6d-b0e6-97d7b918a4e5 service nova] [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] Refreshing network info cache for port fba47d74-c242-4267-a6be-f829d6b2dd6b {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 660.226101] env[63202]: DEBUG nova.network.neutron [-] [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.240531] env[63202]: DEBUG nova.compute.utils [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 660.242311] env[63202]: DEBUG nova.compute.manager [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 660.245578] env[63202]: DEBUG nova.network.neutron [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 660.326317] env[63202]: ERROR nova.compute.manager [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fba47d74-c242-4267-a6be-f829d6b2dd6b, please check neutron logs for more information. [ 660.326317] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 660.326317] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 660.326317] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 660.326317] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 660.326317] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 660.326317] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 660.326317] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 660.326317] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.326317] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 660.326317] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.326317] env[63202]: ERROR nova.compute.manager raise self.value [ 660.326317] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 660.326317] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 660.326317] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.326317] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 660.326883] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.326883] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 660.326883] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fba47d74-c242-4267-a6be-f829d6b2dd6b, please check neutron logs for more information. [ 660.326883] env[63202]: ERROR nova.compute.manager [ 660.326883] env[63202]: Traceback (most recent call last): [ 660.326883] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 660.326883] env[63202]: listener.cb(fileno) [ 660.326883] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 660.326883] env[63202]: result = function(*args, **kwargs) [ 660.326883] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 660.326883] env[63202]: return func(*args, **kwargs) [ 660.326883] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 660.326883] env[63202]: raise e [ 660.326883] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 660.326883] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 660.326883] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 660.326883] env[63202]: created_port_ids = self._update_ports_for_instance( [ 660.326883] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 660.326883] env[63202]: with excutils.save_and_reraise_exception(): [ 660.326883] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.326883] env[63202]: self.force_reraise() [ 660.326883] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.326883] env[63202]: raise self.value [ 660.326883] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 660.326883] env[63202]: updated_port = self._update_port( [ 660.326883] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.326883] env[63202]: _ensure_no_port_binding_failure(port) [ 660.326883] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.326883] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 660.327519] env[63202]: nova.exception.PortBindingFailed: Binding failed for port fba47d74-c242-4267-a6be-f829d6b2dd6b, please check neutron logs for more information. [ 660.327519] env[63202]: Removing descriptor: 16 [ 660.327566] env[63202]: ERROR nova.compute.manager [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fba47d74-c242-4267-a6be-f829d6b2dd6b, please check neutron logs for more information. [ 660.327566] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] Traceback (most recent call last): [ 660.327566] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 660.327566] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] yield resources [ 660.327566] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 660.327566] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] self.driver.spawn(context, instance, image_meta, [ 660.327566] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 660.327566] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] self._vmops.spawn(context, instance, image_meta, injected_files, [ 660.327566] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 660.327566] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] vm_ref = self.build_virtual_machine(instance, [ 660.327566] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 660.327826] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] vif_infos = vmwarevif.get_vif_info(self._session, [ 660.327826] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 660.327826] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] for vif in network_info: [ 660.327826] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 660.327826] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] return self._sync_wrapper(fn, *args, **kwargs) [ 660.327826] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 660.327826] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] self.wait() [ 660.327826] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 660.327826] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] self[:] = self._gt.wait() [ 660.327826] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 660.327826] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] return self._exit_event.wait() [ 660.327826] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 660.327826] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] result = hub.switch() [ 660.328122] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 660.328122] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] return self.greenlet.switch() [ 660.328122] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 660.328122] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] result = function(*args, **kwargs) [ 660.328122] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 660.328122] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] return func(*args, **kwargs) [ 660.328122] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 660.328122] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] raise e [ 660.328122] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 660.328122] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] nwinfo = self.network_api.allocate_for_instance( [ 660.328122] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 660.328122] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] created_port_ids = self._update_ports_for_instance( [ 660.328122] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 660.328396] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] with excutils.save_and_reraise_exception(): [ 660.328396] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.328396] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] self.force_reraise() [ 660.328396] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.328396] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] raise self.value [ 660.328396] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 660.328396] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] updated_port = self._update_port( [ 660.328396] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.328396] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] _ensure_no_port_binding_failure(port) [ 660.328396] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.328396] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] raise exception.PortBindingFailed(port_id=port['id']) [ 660.328396] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] nova.exception.PortBindingFailed: Binding failed for port fba47d74-c242-4267-a6be-f829d6b2dd6b, please check neutron logs for more information. [ 660.328396] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] [ 660.328783] env[63202]: INFO nova.compute.manager [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] Terminating instance [ 660.333038] env[63202]: DEBUG nova.policy [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1236175893a3401cb01b04604168f1e0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '22352c25bca8416a948014391a5389ac', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 660.338792] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] Acquiring lock "refresh_cache-4c84555a-83ac-4896-b40c-cbd7e14eebed" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.639035] env[63202]: DEBUG nova.network.neutron [req-3e696d7b-265d-4b7d-b17f-5acf574b7747 req-346ac9c1-66b7-4f6d-b0e6-97d7b918a4e5 service nova] [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 660.732500] env[63202]: INFO nova.compute.manager [-] [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] Took 1.04 seconds to deallocate network for instance. [ 660.735049] env[63202]: DEBUG nova.compute.claims [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 660.735244] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 660.748993] env[63202]: DEBUG nova.compute.manager [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 660.777019] env[63202]: DEBUG nova.network.neutron [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] Successfully created port: 24fee5ac-16fc-47a2-bba4-a3f2f983a3ca {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 660.792644] env[63202]: DEBUG nova.network.neutron [req-3e696d7b-265d-4b7d-b17f-5acf574b7747 req-346ac9c1-66b7-4f6d-b0e6-97d7b918a4e5 service nova] [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.208753] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e21f2e5-d6f7-474f-8bf4-1740e56db599 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.217144] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd58baab-7bd6-4fad-acfd-9b8aeb1e265c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.256014] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5900a14-7db1-442e-b64a-fa7f9e2b31ce {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.264993] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f465a6a-ea50-40d1-b06e-bcd42de732f3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.281495] env[63202]: DEBUG nova.compute.provider_tree [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 661.295036] env[63202]: DEBUG oslo_concurrency.lockutils [req-3e696d7b-265d-4b7d-b17f-5acf574b7747 req-346ac9c1-66b7-4f6d-b0e6-97d7b918a4e5 service nova] Releasing lock "refresh_cache-4c84555a-83ac-4896-b40c-cbd7e14eebed" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 661.295436] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] Acquired lock "refresh_cache-4c84555a-83ac-4896-b40c-cbd7e14eebed" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 661.295618] env[63202]: DEBUG nova.network.neutron [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 661.762532] env[63202]: DEBUG nova.compute.manager [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 661.784195] env[63202]: DEBUG nova.scheduler.client.report [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 661.792342] env[63202]: DEBUG nova.virt.hardware [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 661.792342] env[63202]: DEBUG nova.virt.hardware [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 661.792342] env[63202]: DEBUG nova.virt.hardware [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 661.792537] env[63202]: DEBUG nova.virt.hardware [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 661.792537] env[63202]: DEBUG nova.virt.hardware [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 661.792537] env[63202]: DEBUG nova.virt.hardware [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 661.792537] env[63202]: DEBUG nova.virt.hardware [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 661.792537] env[63202]: DEBUG nova.virt.hardware [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 661.792676] env[63202]: DEBUG nova.virt.hardware [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 661.792676] env[63202]: DEBUG nova.virt.hardware [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 661.792676] env[63202]: DEBUG nova.virt.hardware [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 661.793332] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8631f7b2-3abf-4841-9403-c18ec2a33a39 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.803212] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95b7b0eb-8d82-4421-af0c-039adf9d4298 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.826039] env[63202]: DEBUG nova.network.neutron [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 661.927880] env[63202]: DEBUG nova.network.neutron [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.195911] env[63202]: DEBUG nova.compute.manager [req-64edb5ed-f631-494e-a357-a035b57fda94 req-87d7415d-6b63-47e1-bb5f-0920efe2b0ff service nova] [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] Received event network-vif-deleted-fba47d74-c242-4267-a6be-f829d6b2dd6b {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 662.299916] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.559s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 662.299916] env[63202]: DEBUG nova.compute.manager [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 662.301504] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.119s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 662.430336] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] Releasing lock "refresh_cache-4c84555a-83ac-4896-b40c-cbd7e14eebed" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 662.432178] env[63202]: DEBUG nova.compute.manager [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 662.432178] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 662.432178] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d54fe245-a052-48a2-87ea-2007242a75e4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.442635] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83771c77-1a99-4d02-b2b4-2ceca0b51900 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.476111] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4c84555a-83ac-4896-b40c-cbd7e14eebed could not be found. [ 662.476329] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 662.476513] env[63202]: INFO nova.compute.manager [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] Took 0.05 seconds to destroy the instance on the hypervisor. [ 662.476762] env[63202]: DEBUG oslo.service.loopingcall [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 662.477011] env[63202]: DEBUG nova.compute.manager [-] [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 662.481131] env[63202]: DEBUG nova.network.neutron [-] [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 662.488915] env[63202]: DEBUG nova.compute.manager [req-685e2eec-a924-4c2d-aea7-523833543e68 req-af1df152-d592-4c59-97e5-38c9fbcd4c3f service nova] [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] Received event network-changed-24fee5ac-16fc-47a2-bba4-a3f2f983a3ca {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 662.489107] env[63202]: DEBUG nova.compute.manager [req-685e2eec-a924-4c2d-aea7-523833543e68 req-af1df152-d592-4c59-97e5-38c9fbcd4c3f service nova] [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] Refreshing instance network info cache due to event network-changed-24fee5ac-16fc-47a2-bba4-a3f2f983a3ca. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 662.489311] env[63202]: DEBUG oslo_concurrency.lockutils [req-685e2eec-a924-4c2d-aea7-523833543e68 req-af1df152-d592-4c59-97e5-38c9fbcd4c3f service nova] Acquiring lock "refresh_cache-45fce1e9-b1bb-48aa-9752-cbd011136abd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 662.491650] env[63202]: DEBUG oslo_concurrency.lockutils [req-685e2eec-a924-4c2d-aea7-523833543e68 req-af1df152-d592-4c59-97e5-38c9fbcd4c3f service nova] Acquired lock "refresh_cache-45fce1e9-b1bb-48aa-9752-cbd011136abd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 662.491650] env[63202]: DEBUG nova.network.neutron [req-685e2eec-a924-4c2d-aea7-523833543e68 req-af1df152-d592-4c59-97e5-38c9fbcd4c3f service nova] [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] Refreshing network info cache for port 24fee5ac-16fc-47a2-bba4-a3f2f983a3ca {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 662.509714] env[63202]: DEBUG nova.network.neutron [-] [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 662.749677] env[63202]: ERROR nova.compute.manager [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 24fee5ac-16fc-47a2-bba4-a3f2f983a3ca, please check neutron logs for more information. [ 662.749677] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 662.749677] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 662.749677] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 662.749677] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 662.749677] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 662.749677] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 662.749677] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 662.749677] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 662.749677] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 662.749677] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 662.749677] env[63202]: ERROR nova.compute.manager raise self.value [ 662.749677] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 662.749677] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 662.749677] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 662.749677] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 662.750261] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 662.750261] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 662.750261] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 24fee5ac-16fc-47a2-bba4-a3f2f983a3ca, please check neutron logs for more information. [ 662.750261] env[63202]: ERROR nova.compute.manager [ 662.750261] env[63202]: Traceback (most recent call last): [ 662.750261] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 662.750261] env[63202]: listener.cb(fileno) [ 662.750261] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 662.750261] env[63202]: result = function(*args, **kwargs) [ 662.750261] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 662.750261] env[63202]: return func(*args, **kwargs) [ 662.750261] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 662.750261] env[63202]: raise e [ 662.750261] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 662.750261] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 662.750261] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 662.750261] env[63202]: created_port_ids = self._update_ports_for_instance( [ 662.750261] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 662.750261] env[63202]: with excutils.save_and_reraise_exception(): [ 662.750261] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 662.750261] env[63202]: self.force_reraise() [ 662.750261] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 662.750261] env[63202]: raise self.value [ 662.750261] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 662.750261] env[63202]: updated_port = self._update_port( [ 662.750261] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 662.750261] env[63202]: _ensure_no_port_binding_failure(port) [ 662.750261] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 662.750261] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 662.750862] env[63202]: nova.exception.PortBindingFailed: Binding failed for port 24fee5ac-16fc-47a2-bba4-a3f2f983a3ca, please check neutron logs for more information. [ 662.750862] env[63202]: Removing descriptor: 20 [ 662.750862] env[63202]: ERROR nova.compute.manager [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 24fee5ac-16fc-47a2-bba4-a3f2f983a3ca, please check neutron logs for more information. [ 662.750862] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] Traceback (most recent call last): [ 662.750862] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 662.750862] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] yield resources [ 662.750862] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 662.750862] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] self.driver.spawn(context, instance, image_meta, [ 662.750862] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 662.750862] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 662.750862] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 662.750862] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] vm_ref = self.build_virtual_machine(instance, [ 662.751137] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 662.751137] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] vif_infos = vmwarevif.get_vif_info(self._session, [ 662.751137] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 662.751137] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] for vif in network_info: [ 662.751137] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 662.751137] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] return self._sync_wrapper(fn, *args, **kwargs) [ 662.751137] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 662.751137] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] self.wait() [ 662.751137] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 662.751137] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] self[:] = self._gt.wait() [ 662.751137] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 662.751137] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] return self._exit_event.wait() [ 662.751137] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 662.751429] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] result = hub.switch() [ 662.751429] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 662.751429] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] return self.greenlet.switch() [ 662.751429] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 662.751429] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] result = function(*args, **kwargs) [ 662.751429] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 662.751429] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] return func(*args, **kwargs) [ 662.751429] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 662.751429] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] raise e [ 662.751429] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 662.751429] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] nwinfo = self.network_api.allocate_for_instance( [ 662.751429] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 662.751429] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] created_port_ids = self._update_ports_for_instance( [ 662.751740] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 662.751740] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] with excutils.save_and_reraise_exception(): [ 662.751740] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 662.751740] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] self.force_reraise() [ 662.751740] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 662.751740] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] raise self.value [ 662.751740] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 662.751740] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] updated_port = self._update_port( [ 662.751740] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 662.751740] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] _ensure_no_port_binding_failure(port) [ 662.751740] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 662.751740] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] raise exception.PortBindingFailed(port_id=port['id']) [ 662.752098] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] nova.exception.PortBindingFailed: Binding failed for port 24fee5ac-16fc-47a2-bba4-a3f2f983a3ca, please check neutron logs for more information. [ 662.752098] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] [ 662.752098] env[63202]: INFO nova.compute.manager [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] Terminating instance [ 662.753516] env[63202]: DEBUG oslo_concurrency.lockutils [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "refresh_cache-45fce1e9-b1bb-48aa-9752-cbd011136abd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 662.805908] env[63202]: DEBUG nova.compute.utils [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 662.812917] env[63202]: DEBUG nova.compute.manager [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 662.812917] env[63202]: DEBUG nova.network.neutron [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 662.878493] env[63202]: DEBUG nova.policy [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '416d20adf8084dbda340480371b9d086', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '321d400a1dfa498a88ad462c133c6493', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 663.012900] env[63202]: DEBUG nova.network.neutron [-] [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.014453] env[63202]: DEBUG nova.network.neutron [req-685e2eec-a924-4c2d-aea7-523833543e68 req-af1df152-d592-4c59-97e5-38c9fbcd4c3f service nova] [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 663.134023] env[63202]: DEBUG nova.network.neutron [req-685e2eec-a924-4c2d-aea7-523833543e68 req-af1df152-d592-4c59-97e5-38c9fbcd4c3f service nova] [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.148056] env[63202]: DEBUG oslo_concurrency.lockutils [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] Acquiring lock "95d2a2d7-443b-43d6-97ff-fe36fd11002f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 663.149491] env[63202]: DEBUG oslo_concurrency.lockutils [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] Lock "95d2a2d7-443b-43d6-97ff-fe36fd11002f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 663.313830] env[63202]: DEBUG nova.compute.manager [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 663.344649] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49043e75-415a-4f1b-bcc8-eff98b5317d0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.352985] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-404d0e3a-efa9-4ee6-8295-dfa3eae414a6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.358010] env[63202]: DEBUG nova.network.neutron [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] Successfully created port: 6fc2c5e8-8533-4f55-9c73-58ec49bc25fa {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 663.389139] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f0e891d-5690-4548-9015-00e385c764c9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.399759] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6381c206-f6aa-4a88-b4e8-d58dd26df903 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.418058] env[63202]: DEBUG nova.compute.provider_tree [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 663.518427] env[63202]: INFO nova.compute.manager [-] [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] Took 1.04 seconds to deallocate network for instance. [ 663.522764] env[63202]: DEBUG nova.compute.claims [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 663.522961] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 663.648948] env[63202]: DEBUG oslo_concurrency.lockutils [req-685e2eec-a924-4c2d-aea7-523833543e68 req-af1df152-d592-4c59-97e5-38c9fbcd4c3f service nova] Releasing lock "refresh_cache-45fce1e9-b1bb-48aa-9752-cbd011136abd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 663.649449] env[63202]: DEBUG oslo_concurrency.lockutils [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquired lock "refresh_cache-45fce1e9-b1bb-48aa-9752-cbd011136abd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.649635] env[63202]: DEBUG nova.network.neutron [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 663.921209] env[63202]: DEBUG nova.scheduler.client.report [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 664.178561] env[63202]: DEBUG nova.network.neutron [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 664.297821] env[63202]: DEBUG nova.network.neutron [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.312205] env[63202]: ERROR nova.compute.manager [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6fc2c5e8-8533-4f55-9c73-58ec49bc25fa, please check neutron logs for more information. [ 664.312205] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 664.312205] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 664.312205] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 664.312205] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 664.312205] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 664.312205] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 664.312205] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 664.312205] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 664.312205] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 664.312205] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 664.312205] env[63202]: ERROR nova.compute.manager raise self.value [ 664.312205] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 664.312205] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 664.312205] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 664.312205] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 664.312657] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 664.312657] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 664.312657] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6fc2c5e8-8533-4f55-9c73-58ec49bc25fa, please check neutron logs for more information. [ 664.312657] env[63202]: ERROR nova.compute.manager [ 664.312657] env[63202]: Traceback (most recent call last): [ 664.312657] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 664.312657] env[63202]: listener.cb(fileno) [ 664.312657] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 664.312657] env[63202]: result = function(*args, **kwargs) [ 664.312657] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 664.312657] env[63202]: return func(*args, **kwargs) [ 664.312657] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 664.312657] env[63202]: raise e [ 664.312657] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 664.312657] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 664.312657] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 664.312657] env[63202]: created_port_ids = self._update_ports_for_instance( [ 664.312657] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 664.312657] env[63202]: with excutils.save_and_reraise_exception(): [ 664.312657] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 664.312657] env[63202]: self.force_reraise() [ 664.312657] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 664.312657] env[63202]: raise self.value [ 664.312657] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 664.312657] env[63202]: updated_port = self._update_port( [ 664.312657] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 664.312657] env[63202]: _ensure_no_port_binding_failure(port) [ 664.312657] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 664.312657] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 664.313452] env[63202]: nova.exception.PortBindingFailed: Binding failed for port 6fc2c5e8-8533-4f55-9c73-58ec49bc25fa, please check neutron logs for more information. [ 664.313452] env[63202]: Removing descriptor: 20 [ 664.326808] env[63202]: DEBUG nova.compute.manager [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 664.351064] env[63202]: DEBUG nova.virt.hardware [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 664.351312] env[63202]: DEBUG nova.virt.hardware [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 664.351462] env[63202]: DEBUG nova.virt.hardware [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 664.351634] env[63202]: DEBUG nova.virt.hardware [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 664.351772] env[63202]: DEBUG nova.virt.hardware [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 664.351914] env[63202]: DEBUG nova.virt.hardware [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 664.352240] env[63202]: DEBUG nova.virt.hardware [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 664.352465] env[63202]: DEBUG nova.virt.hardware [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 664.352690] env[63202]: DEBUG nova.virt.hardware [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 664.352906] env[63202]: DEBUG nova.virt.hardware [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 664.353164] env[63202]: DEBUG nova.virt.hardware [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 664.354277] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-031eba0c-40b5-43fe-b4ab-e7aed3a9d748 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.363208] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64adc5ed-b51f-45ff-9583-d00efc7d9b0f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.378053] env[63202]: ERROR nova.compute.manager [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6fc2c5e8-8533-4f55-9c73-58ec49bc25fa, please check neutron logs for more information. [ 664.378053] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] Traceback (most recent call last): [ 664.378053] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 664.378053] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] yield resources [ 664.378053] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 664.378053] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] self.driver.spawn(context, instance, image_meta, [ 664.378053] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 664.378053] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] self._vmops.spawn(context, instance, image_meta, injected_files, [ 664.378053] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 664.378053] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] vm_ref = self.build_virtual_machine(instance, [ 664.378053] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 664.378429] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] vif_infos = vmwarevif.get_vif_info(self._session, [ 664.378429] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 664.378429] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] for vif in network_info: [ 664.378429] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 664.378429] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] return self._sync_wrapper(fn, *args, **kwargs) [ 664.378429] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 664.378429] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] self.wait() [ 664.378429] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 664.378429] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] self[:] = self._gt.wait() [ 664.378429] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 664.378429] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] return self._exit_event.wait() [ 664.378429] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 664.378429] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] current.throw(*self._exc) [ 664.378890] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 664.378890] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] result = function(*args, **kwargs) [ 664.378890] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 664.378890] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] return func(*args, **kwargs) [ 664.378890] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 664.378890] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] raise e [ 664.378890] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 664.378890] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] nwinfo = self.network_api.allocate_for_instance( [ 664.378890] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 664.378890] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] created_port_ids = self._update_ports_for_instance( [ 664.378890] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 664.378890] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] with excutils.save_and_reraise_exception(): [ 664.378890] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 664.379337] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] self.force_reraise() [ 664.379337] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 664.379337] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] raise self.value [ 664.379337] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 664.379337] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] updated_port = self._update_port( [ 664.379337] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 664.379337] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] _ensure_no_port_binding_failure(port) [ 664.379337] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 664.379337] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] raise exception.PortBindingFailed(port_id=port['id']) [ 664.379337] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] nova.exception.PortBindingFailed: Binding failed for port 6fc2c5e8-8533-4f55-9c73-58ec49bc25fa, please check neutron logs for more information. [ 664.379337] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] [ 664.379337] env[63202]: INFO nova.compute.manager [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] Terminating instance [ 664.380483] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] Acquiring lock "refresh_cache-ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 664.380639] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] Acquired lock "refresh_cache-ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 664.380799] env[63202]: DEBUG nova.network.neutron [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 664.428137] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.127s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 664.428740] env[63202]: ERROR nova.compute.manager [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 740b10d8-5849-4e37-91ad-ad3d7b185b16, please check neutron logs for more information. [ 664.428740] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] Traceback (most recent call last): [ 664.428740] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 664.428740] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] self.driver.spawn(context, instance, image_meta, [ 664.428740] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 664.428740] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] self._vmops.spawn(context, instance, image_meta, injected_files, [ 664.428740] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 664.428740] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] vm_ref = self.build_virtual_machine(instance, [ 664.428740] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 664.428740] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] vif_infos = vmwarevif.get_vif_info(self._session, [ 664.428740] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 664.429099] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] for vif in network_info: [ 664.429099] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 664.429099] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] return self._sync_wrapper(fn, *args, **kwargs) [ 664.429099] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 664.429099] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] self.wait() [ 664.429099] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 664.429099] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] self[:] = self._gt.wait() [ 664.429099] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 664.429099] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] return self._exit_event.wait() [ 664.429099] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 664.429099] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] result = hub.switch() [ 664.429099] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 664.429099] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] return self.greenlet.switch() [ 664.429445] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 664.429445] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] result = function(*args, **kwargs) [ 664.429445] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 664.429445] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] return func(*args, **kwargs) [ 664.429445] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 664.429445] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] raise e [ 664.429445] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 664.429445] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] nwinfo = self.network_api.allocate_for_instance( [ 664.429445] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 664.429445] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] created_port_ids = self._update_ports_for_instance( [ 664.429445] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 664.429445] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] with excutils.save_and_reraise_exception(): [ 664.429445] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 664.429780] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] self.force_reraise() [ 664.429780] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 664.429780] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] raise self.value [ 664.429780] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 664.429780] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] updated_port = self._update_port( [ 664.429780] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 664.429780] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] _ensure_no_port_binding_failure(port) [ 664.429780] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 664.429780] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] raise exception.PortBindingFailed(port_id=port['id']) [ 664.429780] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] nova.exception.PortBindingFailed: Binding failed for port 740b10d8-5849-4e37-91ad-ad3d7b185b16, please check neutron logs for more information. [ 664.429780] env[63202]: ERROR nova.compute.manager [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] [ 664.430083] env[63202]: DEBUG nova.compute.utils [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] Binding failed for port 740b10d8-5849-4e37-91ad-ad3d7b185b16, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 664.430532] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.281s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 664.431972] env[63202]: INFO nova.compute.claims [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: 8cd0415d-dab9-4554-966c-1ea8d822deff] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 664.434782] env[63202]: DEBUG nova.compute.manager [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] Build of instance 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20 was re-scheduled: Binding failed for port 740b10d8-5849-4e37-91ad-ad3d7b185b16, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 664.435213] env[63202]: DEBUG nova.compute.manager [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 664.435426] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquiring lock "refresh_cache-13e91ad9-7fe6-4bbe-8a67-d63a48bfac20" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 664.435567] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquired lock "refresh_cache-13e91ad9-7fe6-4bbe-8a67-d63a48bfac20" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 664.435725] env[63202]: DEBUG nova.network.neutron [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 664.522195] env[63202]: DEBUG nova.compute.manager [req-bb90ea43-0ff5-4689-b597-5caec64e55c2 req-e285856f-74b8-446f-b6bc-c8a599a2ec19 service nova] [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] Received event network-vif-deleted-24fee5ac-16fc-47a2-bba4-a3f2f983a3ca {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 664.522552] env[63202]: DEBUG nova.compute.manager [req-bb90ea43-0ff5-4689-b597-5caec64e55c2 req-e285856f-74b8-446f-b6bc-c8a599a2ec19 service nova] [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] Received event network-changed-6fc2c5e8-8533-4f55-9c73-58ec49bc25fa {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 664.522552] env[63202]: DEBUG nova.compute.manager [req-bb90ea43-0ff5-4689-b597-5caec64e55c2 req-e285856f-74b8-446f-b6bc-c8a599a2ec19 service nova] [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] Refreshing instance network info cache due to event network-changed-6fc2c5e8-8533-4f55-9c73-58ec49bc25fa. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 664.522755] env[63202]: DEBUG oslo_concurrency.lockutils [req-bb90ea43-0ff5-4689-b597-5caec64e55c2 req-e285856f-74b8-446f-b6bc-c8a599a2ec19 service nova] Acquiring lock "refresh_cache-ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 664.802494] env[63202]: DEBUG oslo_concurrency.lockutils [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Releasing lock "refresh_cache-45fce1e9-b1bb-48aa-9752-cbd011136abd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 664.802930] env[63202]: DEBUG nova.compute.manager [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 664.803167] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 664.803470] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3e841d72-0aa7-4e8f-a40d-9d4c4ad3c4b6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.814510] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71332be2-e58e-4bb7-846a-f00065591084 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.839991] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 45fce1e9-b1bb-48aa-9752-cbd011136abd could not be found. [ 664.840254] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 664.840476] env[63202]: INFO nova.compute.manager [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] Took 0.04 seconds to destroy the instance on the hypervisor. [ 664.840686] env[63202]: DEBUG oslo.service.loopingcall [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 664.840892] env[63202]: DEBUG nova.compute.manager [-] [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 664.840988] env[63202]: DEBUG nova.network.neutron [-] [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 664.855601] env[63202]: DEBUG nova.network.neutron [-] [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 664.897459] env[63202]: DEBUG nova.network.neutron [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 664.963117] env[63202]: DEBUG nova.network.neutron [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 664.969990] env[63202]: DEBUG nova.network.neutron [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.056950] env[63202]: DEBUG nova.network.neutron [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.359456] env[63202]: DEBUG nova.network.neutron [-] [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.473652] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] Releasing lock "refresh_cache-ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 665.474152] env[63202]: DEBUG nova.compute.manager [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 665.474315] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 665.474619] env[63202]: DEBUG oslo_concurrency.lockutils [req-bb90ea43-0ff5-4689-b597-5caec64e55c2 req-e285856f-74b8-446f-b6bc-c8a599a2ec19 service nova] Acquired lock "refresh_cache-ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 665.474789] env[63202]: DEBUG nova.network.neutron [req-bb90ea43-0ff5-4689-b597-5caec64e55c2 req-e285856f-74b8-446f-b6bc-c8a599a2ec19 service nova] [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] Refreshing network info cache for port 6fc2c5e8-8533-4f55-9c73-58ec49bc25fa {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 665.476158] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8e051c3d-8b38-4b9c-9a0b-56a67cf24fff {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.488381] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1baf7d8-003c-402f-a59a-24d4c1c4d5df {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.515041] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93 could not be found. [ 665.515139] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 665.515312] env[63202]: INFO nova.compute.manager [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] Took 0.04 seconds to destroy the instance on the hypervisor. [ 665.515545] env[63202]: DEBUG oslo.service.loopingcall [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 665.518090] env[63202]: DEBUG nova.compute.manager [-] [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 665.518185] env[63202]: DEBUG nova.network.neutron [-] [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 665.539600] env[63202]: DEBUG nova.network.neutron [-] [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 665.563196] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Releasing lock "refresh_cache-13e91ad9-7fe6-4bbe-8a67-d63a48bfac20" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 665.563196] env[63202]: DEBUG nova.compute.manager [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 665.563196] env[63202]: DEBUG nova.compute.manager [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 665.563196] env[63202]: DEBUG nova.network.neutron [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 665.579187] env[63202]: DEBUG nova.network.neutron [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 665.840934] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48675c2b-4959-4d5a-8de8-f9fad3fe30c0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.848791] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eae47bb-5376-4c9a-8e88-135ff178d62b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.879719] env[63202]: INFO nova.compute.manager [-] [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] Took 1.04 seconds to deallocate network for instance. [ 665.883178] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4be51087-c148-4b94-a689-76321253b053 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.884821] env[63202]: DEBUG nova.compute.claims [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 665.885042] env[63202]: DEBUG oslo_concurrency.lockutils [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 665.891048] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c56502e8-3a60-4c37-b30e-554a9cd317b8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.904592] env[63202]: DEBUG nova.compute.provider_tree [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 665.992558] env[63202]: DEBUG nova.network.neutron [req-bb90ea43-0ff5-4689-b597-5caec64e55c2 req-e285856f-74b8-446f-b6bc-c8a599a2ec19 service nova] [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 666.045119] env[63202]: DEBUG nova.network.neutron [-] [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.081925] env[63202]: DEBUG nova.network.neutron [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.259678] env[63202]: DEBUG nova.network.neutron [req-bb90ea43-0ff5-4689-b597-5caec64e55c2 req-e285856f-74b8-446f-b6bc-c8a599a2ec19 service nova] [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.407698] env[63202]: DEBUG nova.scheduler.client.report [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 666.547033] env[63202]: INFO nova.compute.manager [-] [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] Took 1.03 seconds to deallocate network for instance. [ 666.549177] env[63202]: DEBUG nova.compute.claims [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 666.549351] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 666.584878] env[63202]: INFO nova.compute.manager [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20] Took 1.02 seconds to deallocate network for instance. [ 666.760193] env[63202]: DEBUG oslo_concurrency.lockutils [req-bb90ea43-0ff5-4689-b597-5caec64e55c2 req-e285856f-74b8-446f-b6bc-c8a599a2ec19 service nova] Releasing lock "refresh_cache-ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 666.760193] env[63202]: DEBUG nova.compute.manager [req-bb90ea43-0ff5-4689-b597-5caec64e55c2 req-e285856f-74b8-446f-b6bc-c8a599a2ec19 service nova] [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] Received event network-vif-deleted-6fc2c5e8-8533-4f55-9c73-58ec49bc25fa {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 666.913046] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.482s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 666.913641] env[63202]: DEBUG nova.compute.manager [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: 8cd0415d-dab9-4554-966c-1ea8d822deff] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 666.916557] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.678s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 666.918102] env[63202]: INFO nova.compute.claims [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 667.422379] env[63202]: DEBUG nova.compute.utils [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 667.425485] env[63202]: DEBUG nova.compute.manager [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: 8cd0415d-dab9-4554-966c-1ea8d822deff] Not allocating networking since 'none' was specified. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 667.617434] env[63202]: INFO nova.scheduler.client.report [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Deleted allocations for instance 13e91ad9-7fe6-4bbe-8a67-d63a48bfac20 [ 667.927174] env[63202]: DEBUG nova.compute.manager [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: 8cd0415d-dab9-4554-966c-1ea8d822deff] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 668.124867] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ef302cb8-340b-44e5-ae9a-5a9a302278fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lock "13e91ad9-7fe6-4bbe-8a67-d63a48bfac20" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 65.631s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 668.264037] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-267e1d83-7043-43e5-aaa8-05429794c942 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.271615] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e0cd331-2551-405b-802c-0f512c382b3e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.303233] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e985926-1908-4134-88fd-043a7b0e6c17 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.310602] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcc62706-e6aa-4af9-8c61-846283b5bd5d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.323454] env[63202]: DEBUG nova.compute.provider_tree [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 668.628624] env[63202]: DEBUG nova.compute.manager [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 668.827918] env[63202]: DEBUG nova.scheduler.client.report [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 668.940550] env[63202]: DEBUG nova.compute.manager [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: 8cd0415d-dab9-4554-966c-1ea8d822deff] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 668.963321] env[63202]: DEBUG nova.virt.hardware [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 668.963588] env[63202]: DEBUG nova.virt.hardware [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 668.963716] env[63202]: DEBUG nova.virt.hardware [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 668.963888] env[63202]: DEBUG nova.virt.hardware [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 668.964357] env[63202]: DEBUG nova.virt.hardware [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 668.964610] env[63202]: DEBUG nova.virt.hardware [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 668.964881] env[63202]: DEBUG nova.virt.hardware [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 668.965072] env[63202]: DEBUG nova.virt.hardware [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 668.965505] env[63202]: DEBUG nova.virt.hardware [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 668.965697] env[63202]: DEBUG nova.virt.hardware [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 668.966013] env[63202]: DEBUG nova.virt.hardware [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 668.966819] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acb7b788-c68b-47d0-9d71-70ba189b8b91 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.975134] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e58ae740-ab88-435e-bbc6-78ab9fef02f0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.988517] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: 8cd0415d-dab9-4554-966c-1ea8d822deff] Instance VIF info [] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 668.994348] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Creating folder: Project (aa78085c2b8f433695c216feacc68722). Parent ref: group-v294090. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 668.994629] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-101998cb-92aa-460b-b9f9-a59a6816db28 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.005622] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Created folder: Project (aa78085c2b8f433695c216feacc68722) in parent group-v294090. [ 669.005836] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Creating folder: Instances. Parent ref: group-v294103. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 669.006157] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e6b944da-94c9-4d93-851b-607435ba7e97 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.014281] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Created folder: Instances in parent group-v294103. [ 669.014679] env[63202]: DEBUG oslo.service.loopingcall [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 669.014679] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8cd0415d-dab9-4554-966c-1ea8d822deff] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 669.014864] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9947308a-9f4b-4b91-92d7-78a19ac5c6cd {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.030248] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 669.030248] env[63202]: value = "task-1385329" [ 669.030248] env[63202]: _type = "Task" [ 669.030248] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 669.037765] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385329, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.156548] env[63202]: DEBUG oslo_concurrency.lockutils [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 669.333968] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.417s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 669.334524] env[63202]: DEBUG nova.compute.manager [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 669.337209] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 18.168s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 669.340372] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 669.340372] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63202) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 669.340372] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.729s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 669.341579] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41d45b9a-8cba-4ef6-8124-e0d2ac459625 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.350087] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca405368-2ab1-429a-971d-0f31b85e4268 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.364768] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c4f3b10-7ff9-4ddf-a062-51c8ebfa58e2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.372170] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc162111-ce44-4dca-8e33-24118f7890a2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.402376] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181096MB free_disk=176GB free_vcpus=48 pci_devices=None {{(pid=63202) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 669.402528] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 669.541180] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385329, 'name': CreateVM_Task} progress is 99%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.756270] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquiring lock "39d9b91b-c009-40e1-ad2b-ebef650188a4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 669.756536] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lock "39d9b91b-c009-40e1-ad2b-ebef650188a4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 669.843985] env[63202]: DEBUG nova.compute.utils [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 669.844618] env[63202]: DEBUG nova.compute.manager [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Not allocating networking since 'none' was specified. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 670.043289] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385329, 'name': CreateVM_Task} progress is 99%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.193852] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fefd567-6601-466c-bc37-b3f981e11bb4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.201576] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5afe295b-5713-4d79-8d8f-0fc8990f0219 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.230857] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-560e907f-fc77-45fe-9b5a-91ca9ad72b4a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.238139] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baecc477-1824-4a19-9fb2-71300fffb017 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.250848] env[63202]: DEBUG nova.compute.provider_tree [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 670.352294] env[63202]: DEBUG nova.compute.manager [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 670.548390] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385329, 'name': CreateVM_Task, 'duration_secs': 1.26184} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 670.548609] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8cd0415d-dab9-4554-966c-1ea8d822deff] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 670.548869] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 670.549035] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.549346] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 670.549653] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b112b41-cad7-49af-95fa-5a8de24bbfd6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.554387] env[63202]: DEBUG oslo_vmware.api [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Waiting for the task: (returnval){ [ 670.554387] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]526ddd3a-8366-0764-45eb-3dd5ee0de99d" [ 670.554387] env[63202]: _type = "Task" [ 670.554387] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 670.561906] env[63202]: DEBUG oslo_vmware.api [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]526ddd3a-8366-0764-45eb-3dd5ee0de99d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.755051] env[63202]: DEBUG nova.scheduler.client.report [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 671.065028] env[63202]: DEBUG oslo_vmware.api [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]526ddd3a-8366-0764-45eb-3dd5ee0de99d, 'name': SearchDatastore_Task, 'duration_secs': 0.024009} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 671.065328] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 671.065548] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: 8cd0415d-dab9-4554-966c-1ea8d822deff] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 671.065773] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 671.065916] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.066133] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 671.066345] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fb7d0b64-0e24-45e3-a243-bee2bf53867a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.087070] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 671.087249] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 671.087960] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-95d0a827-300b-46f4-a3d2-8d44bce11fde {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.092917] env[63202]: DEBUG oslo_vmware.api [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Waiting for the task: (returnval){ [ 671.092917] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52e52022-ad9c-bd26-9910-b019c238354c" [ 671.092917] env[63202]: _type = "Task" [ 671.092917] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 671.099829] env[63202]: DEBUG oslo_vmware.api [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52e52022-ad9c-bd26-9910-b019c238354c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.260038] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.922s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 671.260470] env[63202]: ERROR nova.compute.manager [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6da85fe8-7cce-4523-a80d-56a4ee26d1af, please check neutron logs for more information. [ 671.260470] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] Traceback (most recent call last): [ 671.260470] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 671.260470] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] self.driver.spawn(context, instance, image_meta, [ 671.260470] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 671.260470] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] self._vmops.spawn(context, instance, image_meta, injected_files, [ 671.260470] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 671.260470] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] vm_ref = self.build_virtual_machine(instance, [ 671.260470] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 671.260470] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] vif_infos = vmwarevif.get_vif_info(self._session, [ 671.260470] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 671.260830] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] for vif in network_info: [ 671.260830] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 671.260830] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] return self._sync_wrapper(fn, *args, **kwargs) [ 671.260830] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 671.260830] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] self.wait() [ 671.260830] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 671.260830] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] self[:] = self._gt.wait() [ 671.260830] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 671.260830] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] return self._exit_event.wait() [ 671.260830] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 671.260830] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] result = hub.switch() [ 671.260830] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 671.260830] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] return self.greenlet.switch() [ 671.261178] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 671.261178] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] result = function(*args, **kwargs) [ 671.261178] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 671.261178] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] return func(*args, **kwargs) [ 671.261178] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 671.261178] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] raise e [ 671.261178] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 671.261178] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] nwinfo = self.network_api.allocate_for_instance( [ 671.261178] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 671.261178] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] created_port_ids = self._update_ports_for_instance( [ 671.261178] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 671.261178] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] with excutils.save_and_reraise_exception(): [ 671.261178] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 671.261514] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] self.force_reraise() [ 671.261514] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 671.261514] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] raise self.value [ 671.261514] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 671.261514] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] updated_port = self._update_port( [ 671.261514] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 671.261514] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] _ensure_no_port_binding_failure(port) [ 671.261514] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 671.261514] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] raise exception.PortBindingFailed(port_id=port['id']) [ 671.261514] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] nova.exception.PortBindingFailed: Binding failed for port 6da85fe8-7cce-4523-a80d-56a4ee26d1af, please check neutron logs for more information. [ 671.261514] env[63202]: ERROR nova.compute.manager [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] [ 671.261816] env[63202]: DEBUG nova.compute.utils [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] Binding failed for port 6da85fe8-7cce-4523-a80d-56a4ee26d1af, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 671.262409] env[63202]: DEBUG oslo_concurrency.lockutils [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.939s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 671.263863] env[63202]: INFO nova.compute.claims [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 671.266557] env[63202]: DEBUG nova.compute.manager [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] Build of instance 03e9f5cc-344c-45c8-abff-aae8db411832 was re-scheduled: Binding failed for port 6da85fe8-7cce-4523-a80d-56a4ee26d1af, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 671.266975] env[63202]: DEBUG nova.compute.manager [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 671.267214] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] Acquiring lock "refresh_cache-03e9f5cc-344c-45c8-abff-aae8db411832" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 671.267355] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] Acquired lock "refresh_cache-03e9f5cc-344c-45c8-abff-aae8db411832" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.267535] env[63202]: DEBUG nova.network.neutron [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 671.372232] env[63202]: DEBUG nova.compute.manager [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 671.395288] env[63202]: DEBUG nova.virt.hardware [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 671.395543] env[63202]: DEBUG nova.virt.hardware [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 671.395691] env[63202]: DEBUG nova.virt.hardware [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 671.395906] env[63202]: DEBUG nova.virt.hardware [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 671.396069] env[63202]: DEBUG nova.virt.hardware [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 671.396219] env[63202]: DEBUG nova.virt.hardware [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 671.396422] env[63202]: DEBUG nova.virt.hardware [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 671.396573] env[63202]: DEBUG nova.virt.hardware [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 671.396731] env[63202]: DEBUG nova.virt.hardware [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 671.396889] env[63202]: DEBUG nova.virt.hardware [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 671.397073] env[63202]: DEBUG nova.virt.hardware [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 671.397937] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afcc89dd-2d55-4417-8af5-d213b0667ec7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.405764] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f53eb917-8803-45df-b92f-1171ddfd33ec {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.418754] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Instance VIF info [] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 671.424138] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Creating folder: Project (1b27d72e1c4a437182e5a5351887b7bc). Parent ref: group-v294090. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 671.424374] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6f2292e7-16fa-499f-9ba6-c6be447c36eb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.433188] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Created folder: Project (1b27d72e1c4a437182e5a5351887b7bc) in parent group-v294090. [ 671.433356] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Creating folder: Instances. Parent ref: group-v294106. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 671.433548] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-eb9cc9ae-2daa-4080-b164-5d4ef7985beb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.441302] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Created folder: Instances in parent group-v294106. [ 671.441505] env[63202]: DEBUG oslo.service.loopingcall [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 671.441670] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 671.441844] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c9ae944d-c259-4eb9-9f6f-a32f947b82b6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.457315] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 671.457315] env[63202]: value = "task-1385332" [ 671.457315] env[63202]: _type = "Task" [ 671.457315] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 671.463986] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385332, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.603334] env[63202]: DEBUG oslo_vmware.api [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52e52022-ad9c-bd26-9910-b019c238354c, 'name': SearchDatastore_Task, 'duration_secs': 0.00786} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 671.604128] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-408a77e4-6533-4ee2-8d94-a247c6d8b4e1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.608960] env[63202]: DEBUG oslo_vmware.api [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Waiting for the task: (returnval){ [ 671.608960] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]529712dc-d8ba-d4e9-6d4d-ed5e15064339" [ 671.608960] env[63202]: _type = "Task" [ 671.608960] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 671.616277] env[63202]: DEBUG oslo_vmware.api [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]529712dc-d8ba-d4e9-6d4d-ed5e15064339, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.787426] env[63202]: DEBUG nova.network.neutron [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 671.920354] env[63202]: DEBUG nova.network.neutron [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.967530] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385332, 'name': CreateVM_Task} progress is 99%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.124183] env[63202]: DEBUG oslo_vmware.api [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]529712dc-d8ba-d4e9-6d4d-ed5e15064339, 'name': SearchDatastore_Task, 'duration_secs': 0.103078} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 672.124628] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 672.125033] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] 8cd0415d-dab9-4554-966c-1ea8d822deff/8cd0415d-dab9-4554-966c-1ea8d822deff.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 672.125393] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b75260bc-8b18-40a6-a822-9299687cf25c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.133307] env[63202]: DEBUG oslo_vmware.api [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Waiting for the task: (returnval){ [ 672.133307] env[63202]: value = "task-1385333" [ 672.133307] env[63202]: _type = "Task" [ 672.133307] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 672.144058] env[63202]: DEBUG oslo_vmware.api [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385333, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.425824] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] Releasing lock "refresh_cache-03e9f5cc-344c-45c8-abff-aae8db411832" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 672.426115] env[63202]: DEBUG nova.compute.manager [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 672.426301] env[63202]: DEBUG nova.compute.manager [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 672.427034] env[63202]: DEBUG nova.network.neutron [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 672.468674] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385332, 'name': CreateVM_Task, 'duration_secs': 0.514037} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 672.468856] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 672.469347] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 672.469506] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 672.470814] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 672.472817] env[63202]: DEBUG nova.network.neutron [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 672.473721] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5e53c55f-a445-48e4-a767-73fb76fb376c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.480351] env[63202]: DEBUG oslo_vmware.api [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Waiting for the task: (returnval){ [ 672.480351] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]521904b9-cf8e-7216-b046-7664f9b9b4bb" [ 672.480351] env[63202]: _type = "Task" [ 672.480351] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 672.489356] env[63202]: DEBUG oslo_vmware.api [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]521904b9-cf8e-7216-b046-7664f9b9b4bb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.643872] env[63202]: DEBUG oslo_vmware.api [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385333, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.503361} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 672.646824] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] 8cd0415d-dab9-4554-966c-1ea8d822deff/8cd0415d-dab9-4554-966c-1ea8d822deff.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 672.647093] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: 8cd0415d-dab9-4554-966c-1ea8d822deff] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 672.647812] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-eacced2d-9449-4ede-b1e3-de3648a921ec {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.654310] env[63202]: DEBUG oslo_vmware.api [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Waiting for the task: (returnval){ [ 672.654310] env[63202]: value = "task-1385334" [ 672.654310] env[63202]: _type = "Task" [ 672.654310] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 672.664949] env[63202]: DEBUG oslo_vmware.api [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385334, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.704329] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96b8f7ac-c7d0-4c06-9b9a-1e4cbb1828ed {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.711189] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0c23b60-e2ba-47dd-a44a-eba80650f23d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.741126] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab650cfe-956d-422e-9f34-82cd2ee75bd8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.748582] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cb2b464-ea5a-4716-a86e-612015b93e0c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.761470] env[63202]: DEBUG nova.compute.provider_tree [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 672.976942] env[63202]: DEBUG nova.network.neutron [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.991137] env[63202]: DEBUG oslo_vmware.api [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]521904b9-cf8e-7216-b046-7664f9b9b4bb, 'name': SearchDatastore_Task, 'duration_secs': 0.059695} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 672.991402] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 672.991620] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 672.991847] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 672.992072] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 672.992257] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 672.992502] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e5693b85-74cb-49c3-b8ac-0f3e7ab60521 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.000422] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 673.000592] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 673.001292] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1001b990-dedf-4c44-9c3f-653837a5fd43 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.006041] env[63202]: DEBUG oslo_vmware.api [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Waiting for the task: (returnval){ [ 673.006041] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52f3af31-6d60-6a12-dd6f-7752d83bec9e" [ 673.006041] env[63202]: _type = "Task" [ 673.006041] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.013953] env[63202]: DEBUG oslo_vmware.api [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52f3af31-6d60-6a12-dd6f-7752d83bec9e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.164043] env[63202]: DEBUG oslo_vmware.api [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385334, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.095745} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 673.164043] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: 8cd0415d-dab9-4554-966c-1ea8d822deff] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 673.164553] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a90b9bd-5dad-4c6b-a2b5-66d76002a867 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.183508] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: 8cd0415d-dab9-4554-966c-1ea8d822deff] Reconfiguring VM instance instance-00000019 to attach disk [datastore1] 8cd0415d-dab9-4554-966c-1ea8d822deff/8cd0415d-dab9-4554-966c-1ea8d822deff.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 673.183755] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-244f7584-0b0c-4850-a308-3fd378be8c5e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.202434] env[63202]: DEBUG oslo_vmware.api [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Waiting for the task: (returnval){ [ 673.202434] env[63202]: value = "task-1385335" [ 673.202434] env[63202]: _type = "Task" [ 673.202434] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.209690] env[63202]: DEBUG oslo_vmware.api [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385335, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.264619] env[63202]: DEBUG nova.scheduler.client.report [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 673.479922] env[63202]: INFO nova.compute.manager [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] [instance: 03e9f5cc-344c-45c8-abff-aae8db411832] Took 1.05 seconds to deallocate network for instance. [ 673.515680] env[63202]: DEBUG oslo_vmware.api [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52f3af31-6d60-6a12-dd6f-7752d83bec9e, 'name': SearchDatastore_Task, 'duration_secs': 0.008326} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 673.517128] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0858c3cb-754c-48b1-9952-50484f3905b4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.522754] env[63202]: DEBUG oslo_vmware.api [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Waiting for the task: (returnval){ [ 673.522754] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52f99cf7-c9dd-f677-d3a2-17e5f2d75b2b" [ 673.522754] env[63202]: _type = "Task" [ 673.522754] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.530273] env[63202]: DEBUG oslo_vmware.api [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52f99cf7-c9dd-f677-d3a2-17e5f2d75b2b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.713881] env[63202]: DEBUG oslo_vmware.api [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385335, 'name': ReconfigVM_Task, 'duration_secs': 0.312417} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 673.714130] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: 8cd0415d-dab9-4554-966c-1ea8d822deff] Reconfigured VM instance instance-00000019 to attach disk [datastore1] 8cd0415d-dab9-4554-966c-1ea8d822deff/8cd0415d-dab9-4554-966c-1ea8d822deff.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 673.714946] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3110eab7-1093-4ec0-9d5a-8fcb06841681 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.722264] env[63202]: DEBUG oslo_vmware.api [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Waiting for the task: (returnval){ [ 673.722264] env[63202]: value = "task-1385336" [ 673.722264] env[63202]: _type = "Task" [ 673.722264] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.730437] env[63202]: DEBUG oslo_vmware.api [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385336, 'name': Rename_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.770443] env[63202]: DEBUG oslo_concurrency.lockutils [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.508s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 673.771034] env[63202]: DEBUG nova.compute.manager [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 673.773657] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.722s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 674.033543] env[63202]: DEBUG oslo_vmware.api [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52f99cf7-c9dd-f677-d3a2-17e5f2d75b2b, 'name': SearchDatastore_Task, 'duration_secs': 0.008193} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.033798] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 674.034054] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] e9b604fb-1d88-425c-92b7-008ddaec7c1a/e9b604fb-1d88-425c-92b7-008ddaec7c1a.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 674.034300] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1cf2cd42-be39-4ad9-81c8-aa2a89b0e0a2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.040606] env[63202]: DEBUG oslo_vmware.api [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Waiting for the task: (returnval){ [ 674.040606] env[63202]: value = "task-1385337" [ 674.040606] env[63202]: _type = "Task" [ 674.040606] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.048039] env[63202]: DEBUG oslo_vmware.api [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Task: {'id': task-1385337, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.232648] env[63202]: DEBUG oslo_vmware.api [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385336, 'name': Rename_Task, 'duration_secs': 0.130733} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.232945] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: 8cd0415d-dab9-4554-966c-1ea8d822deff] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 674.233079] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d57e87ae-f081-4b7c-ae7c-19a13dcc9b63 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.239469] env[63202]: DEBUG oslo_vmware.api [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Waiting for the task: (returnval){ [ 674.239469] env[63202]: value = "task-1385338" [ 674.239469] env[63202]: _type = "Task" [ 674.239469] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.247530] env[63202]: DEBUG oslo_vmware.api [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385338, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.278654] env[63202]: DEBUG nova.compute.utils [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 674.283493] env[63202]: DEBUG nova.compute.manager [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Not allocating networking since 'none' was specified. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 674.508191] env[63202]: INFO nova.scheduler.client.report [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] Deleted allocations for instance 03e9f5cc-344c-45c8-abff-aae8db411832 [ 674.552633] env[63202]: DEBUG oslo_vmware.api [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Task: {'id': task-1385337, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.496692} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.552633] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] e9b604fb-1d88-425c-92b7-008ddaec7c1a/e9b604fb-1d88-425c-92b7-008ddaec7c1a.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 674.552633] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 674.552633] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d1c66a50-89df-4bf8-acd4-17e28bbe2a83 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.560934] env[63202]: DEBUG oslo_vmware.api [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Waiting for the task: (returnval){ [ 674.560934] env[63202]: value = "task-1385339" [ 674.560934] env[63202]: _type = "Task" [ 674.560934] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.569147] env[63202]: DEBUG oslo_vmware.api [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Task: {'id': task-1385339, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.718106] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-183a4a60-925b-4130-aad5-66840aade662 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.726147] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d8d72ee-269a-4d5e-93c7-8f6ff4bfcdfa {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.760318] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43dc9cfa-5223-4d7c-9a7d-c0274ebf7126 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.770075] env[63202]: DEBUG oslo_vmware.api [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385338, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.771019] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b038a241-7779-4de2-a873-40336f8945fc {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.784104] env[63202]: DEBUG nova.compute.manager [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 674.786800] env[63202]: DEBUG nova.compute.provider_tree [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 675.019605] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b361ddff-3a71-4c56-8559-bd9a12f7a370 tempest-ServersNegativeTestMultiTenantJSON-853169975 tempest-ServersNegativeTestMultiTenantJSON-853169975-project-member] Lock "03e9f5cc-344c-45c8-abff-aae8db411832" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.754s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 675.071439] env[63202]: DEBUG oslo_vmware.api [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Task: {'id': task-1385339, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068341} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.071665] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 675.072454] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1e122a5-2b8d-40e1-b982-4fb99de44e28 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.097500] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Reconfiguring VM instance instance-0000001a to attach disk [datastore1] e9b604fb-1d88-425c-92b7-008ddaec7c1a/e9b604fb-1d88-425c-92b7-008ddaec7c1a.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 675.099024] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a92e1f47-3894-4ff2-b71c-ae7e9fbb26d4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.120161] env[63202]: DEBUG oslo_vmware.api [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Waiting for the task: (returnval){ [ 675.120161] env[63202]: value = "task-1385340" [ 675.120161] env[63202]: _type = "Task" [ 675.120161] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.127845] env[63202]: DEBUG oslo_vmware.api [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Task: {'id': task-1385340, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.266796] env[63202]: DEBUG oslo_vmware.api [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385338, 'name': PowerOnVM_Task, 'duration_secs': 0.627336} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.267140] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: 8cd0415d-dab9-4554-966c-1ea8d822deff] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 675.267400] env[63202]: INFO nova.compute.manager [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: 8cd0415d-dab9-4554-966c-1ea8d822deff] Took 6.33 seconds to spawn the instance on the hypervisor. [ 675.267628] env[63202]: DEBUG nova.compute.manager [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: 8cd0415d-dab9-4554-966c-1ea8d822deff] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 675.268453] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62010e50-5ae9-44d6-b06a-d3d879d6a00f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.295021] env[63202]: DEBUG nova.scheduler.client.report [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 675.524084] env[63202]: DEBUG nova.compute.manager [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 675.630312] env[63202]: DEBUG oslo_vmware.api [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Task: {'id': task-1385340, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.788020] env[63202]: INFO nova.compute.manager [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: 8cd0415d-dab9-4554-966c-1ea8d822deff] Took 28.67 seconds to build instance. [ 675.800102] env[63202]: DEBUG nova.compute.manager [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 675.803788] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.029s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 675.804649] env[63202]: ERROR nova.compute.manager [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port bde67ad1-7993-4c30-9aa9-13ea5c2f1800, please check neutron logs for more information. [ 675.804649] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] Traceback (most recent call last): [ 675.804649] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 675.804649] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] self.driver.spawn(context, instance, image_meta, [ 675.804649] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 675.804649] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 675.804649] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 675.804649] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] vm_ref = self.build_virtual_machine(instance, [ 675.804649] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 675.804649] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] vif_infos = vmwarevif.get_vif_info(self._session, [ 675.804649] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 675.804980] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] for vif in network_info: [ 675.804980] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 675.804980] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] return self._sync_wrapper(fn, *args, **kwargs) [ 675.804980] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 675.804980] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] self.wait() [ 675.804980] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 675.804980] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] self[:] = self._gt.wait() [ 675.804980] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 675.804980] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] return self._exit_event.wait() [ 675.804980] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 675.804980] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] result = hub.switch() [ 675.804980] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 675.804980] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] return self.greenlet.switch() [ 675.805340] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 675.805340] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] result = function(*args, **kwargs) [ 675.805340] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 675.805340] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] return func(*args, **kwargs) [ 675.805340] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 675.805340] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] raise e [ 675.805340] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 675.805340] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] nwinfo = self.network_api.allocate_for_instance( [ 675.805340] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 675.805340] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] created_port_ids = self._update_ports_for_instance( [ 675.805340] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 675.805340] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] with excutils.save_and_reraise_exception(): [ 675.805340] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 675.805690] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] self.force_reraise() [ 675.805690] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 675.805690] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] raise self.value [ 675.805690] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 675.805690] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] updated_port = self._update_port( [ 675.805690] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 675.805690] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] _ensure_no_port_binding_failure(port) [ 675.805690] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 675.805690] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] raise exception.PortBindingFailed(port_id=port['id']) [ 675.805690] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] nova.exception.PortBindingFailed: Binding failed for port bde67ad1-7993-4c30-9aa9-13ea5c2f1800, please check neutron logs for more information. [ 675.805690] env[63202]: ERROR nova.compute.manager [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] [ 675.805987] env[63202]: DEBUG nova.compute.utils [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] Binding failed for port bde67ad1-7993-4c30-9aa9-13ea5c2f1800, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 675.807567] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.072s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 675.810532] env[63202]: DEBUG nova.compute.manager [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] Build of instance ef69de02-6e43-4489-a35a-8a9096e54fbd was re-scheduled: Binding failed for port bde67ad1-7993-4c30-9aa9-13ea5c2f1800, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 675.814452] env[63202]: DEBUG nova.compute.manager [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 675.814452] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] Acquiring lock "refresh_cache-ef69de02-6e43-4489-a35a-8a9096e54fbd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 675.814623] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] Acquired lock "refresh_cache-ef69de02-6e43-4489-a35a-8a9096e54fbd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 675.814784] env[63202]: DEBUG nova.network.neutron [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 675.838016] env[63202]: DEBUG nova.virt.hardware [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 675.838298] env[63202]: DEBUG nova.virt.hardware [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 675.838454] env[63202]: DEBUG nova.virt.hardware [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 675.838629] env[63202]: DEBUG nova.virt.hardware [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 675.838767] env[63202]: DEBUG nova.virt.hardware [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 675.838922] env[63202]: DEBUG nova.virt.hardware [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 675.839258] env[63202]: DEBUG nova.virt.hardware [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 675.839515] env[63202]: DEBUG nova.virt.hardware [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 675.839764] env[63202]: DEBUG nova.virt.hardware [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 675.839949] env[63202]: DEBUG nova.virt.hardware [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 675.840138] env[63202]: DEBUG nova.virt.hardware [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 675.841252] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e2bede1-4a75-48ec-a346-a675f78fafef {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.849410] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58d9b221-4773-444e-abf1-163dd2a32a70 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.863458] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Instance VIF info [] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 675.869037] env[63202]: DEBUG oslo.service.loopingcall [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 675.869523] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 675.869732] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2411d3b3-ca66-4024-8614-94f737c5bb4d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.887805] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 675.887805] env[63202]: value = "task-1385341" [ 675.887805] env[63202]: _type = "Task" [ 675.887805] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.895735] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385341, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.055417] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 676.130728] env[63202]: DEBUG oslo_vmware.api [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Task: {'id': task-1385340, 'name': ReconfigVM_Task, 'duration_secs': 0.982723} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.131046] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Reconfigured VM instance instance-0000001a to attach disk [datastore1] e9b604fb-1d88-425c-92b7-008ddaec7c1a/e9b604fb-1d88-425c-92b7-008ddaec7c1a.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 676.132213] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-35c87f26-dc2f-4d28-9c06-8411d637a772 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.139332] env[63202]: DEBUG oslo_vmware.api [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Waiting for the task: (returnval){ [ 676.139332] env[63202]: value = "task-1385342" [ 676.139332] env[63202]: _type = "Task" [ 676.139332] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.147964] env[63202]: DEBUG oslo_vmware.api [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Task: {'id': task-1385342, 'name': Rename_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.290060] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6d38a357-e0e1-4bc9-b470-8244f620b436 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Lock "8cd0415d-dab9-4554-966c-1ea8d822deff" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 59.868s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 676.344340] env[63202]: DEBUG nova.network.neutron [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 676.398206] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385341, 'name': CreateVM_Task, 'duration_secs': 0.257876} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.400790] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 676.401449] env[63202]: DEBUG oslo_concurrency.lockutils [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 676.401773] env[63202]: DEBUG oslo_concurrency.lockutils [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.402104] env[63202]: DEBUG oslo_concurrency.lockutils [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 676.402423] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f80d487c-b521-4c94-8c8f-7df04b5623b1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.408868] env[63202]: DEBUG oslo_vmware.api [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Waiting for the task: (returnval){ [ 676.408868] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]525025ca-d89b-3dd5-d1f0-252eb46aa28d" [ 676.408868] env[63202]: _type = "Task" [ 676.408868] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.417779] env[63202]: DEBUG oslo_vmware.api [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]525025ca-d89b-3dd5-d1f0-252eb46aa28d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.474701] env[63202]: DEBUG nova.network.neutron [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.651191] env[63202]: DEBUG oslo_vmware.api [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Task: {'id': task-1385342, 'name': Rename_Task, 'duration_secs': 0.141251} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.651463] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 676.651706] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9f40247c-7cc3-4505-a2e0-5a9d4ae9cc04 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.658019] env[63202]: DEBUG oslo_vmware.api [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Waiting for the task: (returnval){ [ 676.658019] env[63202]: value = "task-1385343" [ 676.658019] env[63202]: _type = "Task" [ 676.658019] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.672622] env[63202]: DEBUG oslo_vmware.api [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Task: {'id': task-1385343, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.759451] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-506a6548-c57a-4474-a921-405dff174250 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.767248] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8af8e2b7-8a9c-4063-a738-d4d5f0a1ab8a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.798485] env[63202]: DEBUG nova.compute.manager [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 676.801691] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06daea60-39b8-45c6-a8da-868307132fe7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.809602] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b76f96bd-ff4f-4716-9f44-079268015ee8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.823589] env[63202]: DEBUG nova.compute.provider_tree [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 676.920367] env[63202]: DEBUG oslo_vmware.api [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]525025ca-d89b-3dd5-d1f0-252eb46aa28d, 'name': SearchDatastore_Task, 'duration_secs': 0.013405} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.920367] env[63202]: DEBUG oslo_concurrency.lockutils [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 676.920367] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 676.920367] env[63202]: DEBUG oslo_concurrency.lockutils [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 676.920595] env[63202]: DEBUG oslo_concurrency.lockutils [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.920626] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 676.921335] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a276ee7c-5d78-4031-aa3e-82a4a4fd2d0b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.932406] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 676.933427] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 676.934302] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1ba04df-47bf-4e5e-909b-01837c9f731f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.941607] env[63202]: DEBUG oslo_vmware.api [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Waiting for the task: (returnval){ [ 676.941607] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]525b7494-898c-e28e-8ba4-a8dedb8bfa68" [ 676.941607] env[63202]: _type = "Task" [ 676.941607] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.949471] env[63202]: DEBUG oslo_vmware.api [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]525b7494-898c-e28e-8ba4-a8dedb8bfa68, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.977543] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] Releasing lock "refresh_cache-ef69de02-6e43-4489-a35a-8a9096e54fbd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 676.977765] env[63202]: DEBUG nova.compute.manager [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 676.977976] env[63202]: DEBUG nova.compute.manager [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 676.978171] env[63202]: DEBUG nova.network.neutron [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 676.991512] env[63202]: DEBUG nova.network.neutron [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 677.169466] env[63202]: DEBUG oslo_vmware.api [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Task: {'id': task-1385343, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.324427] env[63202]: DEBUG oslo_concurrency.lockutils [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 677.326491] env[63202]: DEBUG nova.scheduler.client.report [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 677.454209] env[63202]: DEBUG oslo_vmware.api [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]525b7494-898c-e28e-8ba4-a8dedb8bfa68, 'name': SearchDatastore_Task, 'duration_secs': 0.013724} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.455050] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cfd44fb9-71a5-4e6e-b268-5e407244770d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.460350] env[63202]: DEBUG oslo_vmware.api [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Waiting for the task: (returnval){ [ 677.460350] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52a10158-80e2-06af-72e7-be195d589288" [ 677.460350] env[63202]: _type = "Task" [ 677.460350] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.467836] env[63202]: DEBUG oslo_vmware.api [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52a10158-80e2-06af-72e7-be195d589288, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.493959] env[63202]: DEBUG nova.network.neutron [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.670611] env[63202]: DEBUG oslo_vmware.api [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Task: {'id': task-1385343, 'name': PowerOnVM_Task, 'duration_secs': 0.627324} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.670611] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 677.670611] env[63202]: INFO nova.compute.manager [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Took 6.30 seconds to spawn the instance on the hypervisor. [ 677.670611] env[63202]: DEBUG nova.compute.manager [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 677.671173] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4637d8f-db7f-4570-95d8-2f527ebf1e55 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.831938] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.024s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 677.832601] env[63202]: ERROR nova.compute.manager [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3541508a-02bc-425d-9c91-ef407d837b04, please check neutron logs for more information. [ 677.832601] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] Traceback (most recent call last): [ 677.832601] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 677.832601] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] self.driver.spawn(context, instance, image_meta, [ 677.832601] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 677.832601] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 677.832601] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 677.832601] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] vm_ref = self.build_virtual_machine(instance, [ 677.832601] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 677.832601] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] vif_infos = vmwarevif.get_vif_info(self._session, [ 677.832601] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 677.832938] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] for vif in network_info: [ 677.832938] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 677.832938] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] return self._sync_wrapper(fn, *args, **kwargs) [ 677.832938] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 677.832938] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] self.wait() [ 677.832938] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 677.832938] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] self[:] = self._gt.wait() [ 677.832938] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 677.832938] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] return self._exit_event.wait() [ 677.832938] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 677.832938] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] result = hub.switch() [ 677.832938] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 677.832938] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] return self.greenlet.switch() [ 677.833349] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 677.833349] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] result = function(*args, **kwargs) [ 677.833349] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 677.833349] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] return func(*args, **kwargs) [ 677.833349] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 677.833349] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] raise e [ 677.833349] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 677.833349] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] nwinfo = self.network_api.allocate_for_instance( [ 677.833349] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 677.833349] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] created_port_ids = self._update_ports_for_instance( [ 677.833349] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 677.833349] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] with excutils.save_and_reraise_exception(): [ 677.833349] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 677.833709] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] self.force_reraise() [ 677.833709] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 677.833709] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] raise self.value [ 677.833709] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 677.833709] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] updated_port = self._update_port( [ 677.833709] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 677.833709] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] _ensure_no_port_binding_failure(port) [ 677.833709] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 677.833709] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] raise exception.PortBindingFailed(port_id=port['id']) [ 677.833709] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] nova.exception.PortBindingFailed: Binding failed for port 3541508a-02bc-425d-9c91-ef407d837b04, please check neutron logs for more information. [ 677.833709] env[63202]: ERROR nova.compute.manager [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] [ 677.834015] env[63202]: DEBUG nova.compute.utils [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] Binding failed for port 3541508a-02bc-425d-9c91-ef407d837b04, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 677.834509] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.312s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 677.837438] env[63202]: DEBUG nova.compute.manager [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] Build of instance bc8fc812-1e66-4e14-a39c-1c226e7fe2b9 was re-scheduled: Binding failed for port 3541508a-02bc-425d-9c91-ef407d837b04, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 677.837848] env[63202]: DEBUG nova.compute.manager [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 677.838105] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Acquiring lock "refresh_cache-bc8fc812-1e66-4e14-a39c-1c226e7fe2b9" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 677.838255] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Acquired lock "refresh_cache-bc8fc812-1e66-4e14-a39c-1c226e7fe2b9" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 677.838429] env[63202]: DEBUG nova.network.neutron [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 677.971394] env[63202]: DEBUG oslo_vmware.api [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52a10158-80e2-06af-72e7-be195d589288, 'name': SearchDatastore_Task, 'duration_secs': 0.010638} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.971448] env[63202]: DEBUG oslo_concurrency.lockutils [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 677.971658] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] c8d27e19-0f3c-4cc3-826b-e07b3a635e5e/c8d27e19-0f3c-4cc3-826b-e07b3a635e5e.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 677.971897] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d29ab2bd-d931-40af-9394-23b6cc04828a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.980652] env[63202]: DEBUG oslo_vmware.api [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Waiting for the task: (returnval){ [ 677.980652] env[63202]: value = "task-1385344" [ 677.980652] env[63202]: _type = "Task" [ 677.980652] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.987617] env[63202]: DEBUG oslo_vmware.api [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385344, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.996934] env[63202]: INFO nova.compute.manager [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] [instance: ef69de02-6e43-4489-a35a-8a9096e54fbd] Took 1.02 seconds to deallocate network for instance. [ 678.196642] env[63202]: INFO nova.compute.manager [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Took 28.99 seconds to build instance. [ 678.365927] env[63202]: DEBUG nova.network.neutron [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 678.489998] env[63202]: DEBUG oslo_vmware.api [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385344, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.491174] env[63202]: DEBUG nova.network.neutron [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.698439] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c6f3171f-1b51-4a02-bd8c-ff921870bb00 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Lock "e9b604fb-1d88-425c-92b7-008ddaec7c1a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 62.120s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 678.777030] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caf23d29-3b0a-481f-933e-5135c7721f3d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.785644] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82dfd55f-c2fa-4d0b-9e2d-6ac467fefcb3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.816861] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f5ac7b0-c71c-4699-b75a-a5ff519170f2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.824638] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3f78665-9067-4d85-bc9f-702ab8a697ee {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.838105] env[63202]: DEBUG nova.compute.provider_tree [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 678.989510] env[63202]: DEBUG oslo_vmware.api [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385344, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.993124] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Releasing lock "refresh_cache-bc8fc812-1e66-4e14-a39c-1c226e7fe2b9" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 678.993759] env[63202]: DEBUG nova.compute.manager [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 678.993759] env[63202]: DEBUG nova.compute.manager [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 678.993759] env[63202]: DEBUG nova.network.neutron [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 679.010874] env[63202]: DEBUG nova.network.neutron [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 679.026441] env[63202]: INFO nova.scheduler.client.report [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] Deleted allocations for instance ef69de02-6e43-4489-a35a-8a9096e54fbd [ 679.088849] env[63202]: INFO nova.compute.manager [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Rebuilding instance [ 679.134898] env[63202]: DEBUG nova.compute.manager [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 679.135830] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5c0edcb-ccd1-42f3-b408-5aa7923ea2dc {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.201809] env[63202]: DEBUG nova.compute.manager [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] [instance: de5f9555-3959-41cd-b63a-db128a001631] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 679.341126] env[63202]: DEBUG nova.scheduler.client.report [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 679.492487] env[63202]: DEBUG oslo_vmware.api [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385344, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.347281} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.492719] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] c8d27e19-0f3c-4cc3-826b-e07b3a635e5e/c8d27e19-0f3c-4cc3-826b-e07b3a635e5e.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 679.493179] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 679.493260] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-959dacd9-9031-4b72-9735-890766b2dc11 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.500161] env[63202]: DEBUG oslo_vmware.api [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Waiting for the task: (returnval){ [ 679.500161] env[63202]: value = "task-1385345" [ 679.500161] env[63202]: _type = "Task" [ 679.500161] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.507751] env[63202]: DEBUG oslo_vmware.api [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385345, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.513270] env[63202]: DEBUG nova.network.neutron [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.535853] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b0d36060-9e10-49a0-b0f2-41a03a6c55f1 tempest-ServersWithSpecificFlavorTestJSON-1769758871 tempest-ServersWithSpecificFlavorTestJSON-1769758871-project-member] Lock "ef69de02-6e43-4489-a35a-8a9096e54fbd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.271s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 679.648054] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 679.648054] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-06264f25-0aac-45e4-97d9-fa427cc4f252 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.655456] env[63202]: DEBUG oslo_vmware.api [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Waiting for the task: (returnval){ [ 679.655456] env[63202]: value = "task-1385346" [ 679.655456] env[63202]: _type = "Task" [ 679.655456] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.663968] env[63202]: DEBUG oslo_vmware.api [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Task: {'id': task-1385346, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.724927] env[63202]: DEBUG oslo_concurrency.lockutils [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 679.849587] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.015s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 679.850292] env[63202]: ERROR nova.compute.manager [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fba47d74-c242-4267-a6be-f829d6b2dd6b, please check neutron logs for more information. [ 679.850292] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] Traceback (most recent call last): [ 679.850292] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 679.850292] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] self.driver.spawn(context, instance, image_meta, [ 679.850292] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 679.850292] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] self._vmops.spawn(context, instance, image_meta, injected_files, [ 679.850292] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 679.850292] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] vm_ref = self.build_virtual_machine(instance, [ 679.850292] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 679.850292] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] vif_infos = vmwarevif.get_vif_info(self._session, [ 679.850292] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 679.850600] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] for vif in network_info: [ 679.850600] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 679.850600] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] return self._sync_wrapper(fn, *args, **kwargs) [ 679.850600] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 679.850600] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] self.wait() [ 679.850600] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 679.850600] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] self[:] = self._gt.wait() [ 679.850600] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 679.850600] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] return self._exit_event.wait() [ 679.850600] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 679.850600] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] result = hub.switch() [ 679.850600] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 679.850600] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] return self.greenlet.switch() [ 679.850956] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 679.850956] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] result = function(*args, **kwargs) [ 679.850956] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 679.850956] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] return func(*args, **kwargs) [ 679.850956] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 679.850956] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] raise e [ 679.850956] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 679.850956] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] nwinfo = self.network_api.allocate_for_instance( [ 679.850956] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 679.850956] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] created_port_ids = self._update_ports_for_instance( [ 679.850956] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 679.850956] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] with excutils.save_and_reraise_exception(): [ 679.850956] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 679.851329] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] self.force_reraise() [ 679.851329] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 679.851329] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] raise self.value [ 679.851329] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 679.851329] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] updated_port = self._update_port( [ 679.851329] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 679.851329] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] _ensure_no_port_binding_failure(port) [ 679.851329] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 679.851329] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] raise exception.PortBindingFailed(port_id=port['id']) [ 679.851329] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] nova.exception.PortBindingFailed: Binding failed for port fba47d74-c242-4267-a6be-f829d6b2dd6b, please check neutron logs for more information. [ 679.851329] env[63202]: ERROR nova.compute.manager [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] [ 679.851635] env[63202]: DEBUG nova.compute.utils [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] Binding failed for port fba47d74-c242-4267-a6be-f829d6b2dd6b, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 679.852700] env[63202]: DEBUG oslo_concurrency.lockutils [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.968s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 679.856332] env[63202]: DEBUG nova.compute.manager [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] Build of instance 4c84555a-83ac-4896-b40c-cbd7e14eebed was re-scheduled: Binding failed for port fba47d74-c242-4267-a6be-f829d6b2dd6b, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 679.856756] env[63202]: DEBUG nova.compute.manager [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 679.856980] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] Acquiring lock "refresh_cache-4c84555a-83ac-4896-b40c-cbd7e14eebed" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 679.857142] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] Acquired lock "refresh_cache-4c84555a-83ac-4896-b40c-cbd7e14eebed" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.857296] env[63202]: DEBUG nova.network.neutron [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 680.009488] env[63202]: DEBUG oslo_vmware.api [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385345, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061581} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.009754] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 680.010578] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80bf4280-4336-43ae-b581-502a155e6aed {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.022129] env[63202]: INFO nova.compute.manager [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: bc8fc812-1e66-4e14-a39c-1c226e7fe2b9] Took 1.03 seconds to deallocate network for instance. [ 680.034175] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Reconfiguring VM instance instance-0000001b to attach disk [datastore1] c8d27e19-0f3c-4cc3-826b-e07b3a635e5e/c8d27e19-0f3c-4cc3-826b-e07b3a635e5e.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 680.035101] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-44c96de7-17c4-413d-b483-6e52b5f48df7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.049905] env[63202]: DEBUG nova.compute.manager [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] [instance: 2f87939c-d615-4d13-8695-971e83ca7843] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 680.058632] env[63202]: DEBUG oslo_vmware.api [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Waiting for the task: (returnval){ [ 680.058632] env[63202]: value = "task-1385347" [ 680.058632] env[63202]: _type = "Task" [ 680.058632] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.068259] env[63202]: DEBUG oslo_vmware.api [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385347, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.166893] env[63202]: DEBUG oslo_vmware.api [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Task: {'id': task-1385346, 'name': PowerOffVM_Task, 'duration_secs': 0.289172} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.166893] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 680.166893] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 680.167808] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-680b5f71-7924-4c44-9c2c-fa1c9af74025 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.174656] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 680.175049] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ff9df0ae-23e3-404c-917a-2aeafd35e153 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.200040] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 680.200548] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Deleting contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 680.200749] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Deleting the datastore file [datastore1] e9b604fb-1d88-425c-92b7-008ddaec7c1a {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 680.201019] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cf43c101-cb81-412b-9532-f6beee5e6686 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.208909] env[63202]: DEBUG oslo_vmware.api [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Waiting for the task: (returnval){ [ 680.208909] env[63202]: value = "task-1385349" [ 680.208909] env[63202]: _type = "Task" [ 680.208909] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.216470] env[63202]: DEBUG oslo_vmware.api [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Task: {'id': task-1385349, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.386667] env[63202]: DEBUG nova.network.neutron [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 680.507597] env[63202]: DEBUG nova.network.neutron [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.574361] env[63202]: DEBUG oslo_vmware.api [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385347, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.591294] env[63202]: DEBUG oslo_concurrency.lockutils [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 680.721242] env[63202]: DEBUG oslo_vmware.api [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Task: {'id': task-1385349, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.340405} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.721451] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 680.721553] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Deleted contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 680.721698] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 680.799017] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a16f7b1-f8ab-42f2-8837-61a1347f8c20 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.805406] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd1d4428-1b75-429f-a0e4-79af62d92871 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.836034] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa986c84-883a-4810-b162-3ae68575e0e0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.843494] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0897ebb-08a6-4311-a954-2434eb199c74 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.858914] env[63202]: DEBUG nova.compute.provider_tree [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 681.012332] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] Releasing lock "refresh_cache-4c84555a-83ac-4896-b40c-cbd7e14eebed" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 681.012653] env[63202]: DEBUG nova.compute.manager [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 681.012848] env[63202]: DEBUG nova.compute.manager [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 681.013022] env[63202]: DEBUG nova.network.neutron [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 681.032236] env[63202]: DEBUG nova.network.neutron [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 681.062368] env[63202]: INFO nova.scheduler.client.report [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Deleted allocations for instance bc8fc812-1e66-4e14-a39c-1c226e7fe2b9 [ 681.077661] env[63202]: DEBUG oslo_vmware.api [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385347, 'name': ReconfigVM_Task, 'duration_secs': 0.924621} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 681.077931] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Reconfigured VM instance instance-0000001b to attach disk [datastore1] c8d27e19-0f3c-4cc3-826b-e07b3a635e5e/c8d27e19-0f3c-4cc3-826b-e07b3a635e5e.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 681.078589] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-356e6f04-ee8c-4efd-afae-40f0e00e1cca {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.086241] env[63202]: DEBUG oslo_vmware.api [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Waiting for the task: (returnval){ [ 681.086241] env[63202]: value = "task-1385350" [ 681.086241] env[63202]: _type = "Task" [ 681.086241] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.094642] env[63202]: DEBUG oslo_vmware.api [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385350, 'name': Rename_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.361566] env[63202]: DEBUG nova.scheduler.client.report [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 681.541651] env[63202]: DEBUG nova.network.neutron [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.573124] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ebf98e72-8912-44b2-a9b0-e3e10637e925 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Lock "bc8fc812-1e66-4e14-a39c-1c226e7fe2b9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.717s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 681.597608] env[63202]: DEBUG oslo_vmware.api [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385350, 'name': Rename_Task, 'duration_secs': 0.12889} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 681.598100] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 681.598471] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9b8d0a72-149a-4de3-bcd6-2a6197bb49c0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.605510] env[63202]: DEBUG oslo_vmware.api [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Waiting for the task: (returnval){ [ 681.605510] env[63202]: value = "task-1385351" [ 681.605510] env[63202]: _type = "Task" [ 681.605510] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.616170] env[63202]: DEBUG oslo_vmware.api [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385351, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.759037] env[63202]: DEBUG nova.virt.hardware [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 681.759318] env[63202]: DEBUG nova.virt.hardware [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 681.759471] env[63202]: DEBUG nova.virt.hardware [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 681.759647] env[63202]: DEBUG nova.virt.hardware [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 681.759789] env[63202]: DEBUG nova.virt.hardware [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 681.759932] env[63202]: DEBUG nova.virt.hardware [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 681.760403] env[63202]: DEBUG nova.virt.hardware [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 681.760642] env[63202]: DEBUG nova.virt.hardware [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 681.760842] env[63202]: DEBUG nova.virt.hardware [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 681.761042] env[63202]: DEBUG nova.virt.hardware [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 681.761255] env[63202]: DEBUG nova.virt.hardware [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 681.762144] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c731e1cf-b03d-4c02-9eef-b2dc2ac324c3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.770693] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac587d6e-619d-47e2-b331-5bcc1b964bc3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.786095] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Instance VIF info [] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 681.793531] env[63202]: DEBUG oslo.service.loopingcall [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 681.793531] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 681.793531] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6ae2655e-9776-497e-b97f-549dd28b5199 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.810474] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 681.810474] env[63202]: value = "task-1385352" [ 681.810474] env[63202]: _type = "Task" [ 681.810474] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.818351] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385352, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.866047] env[63202]: DEBUG oslo_concurrency.lockutils [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.013s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 681.866930] env[63202]: ERROR nova.compute.manager [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 24fee5ac-16fc-47a2-bba4-a3f2f983a3ca, please check neutron logs for more information. [ 681.866930] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] Traceback (most recent call last): [ 681.866930] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 681.866930] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] self.driver.spawn(context, instance, image_meta, [ 681.866930] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 681.866930] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 681.866930] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 681.866930] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] vm_ref = self.build_virtual_machine(instance, [ 681.866930] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 681.866930] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] vif_infos = vmwarevif.get_vif_info(self._session, [ 681.866930] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 681.867332] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] for vif in network_info: [ 681.867332] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 681.867332] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] return self._sync_wrapper(fn, *args, **kwargs) [ 681.867332] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 681.867332] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] self.wait() [ 681.867332] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 681.867332] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] self[:] = self._gt.wait() [ 681.867332] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 681.867332] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] return self._exit_event.wait() [ 681.867332] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 681.867332] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] result = hub.switch() [ 681.867332] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 681.867332] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] return self.greenlet.switch() [ 681.867766] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 681.867766] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] result = function(*args, **kwargs) [ 681.867766] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 681.867766] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] return func(*args, **kwargs) [ 681.867766] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 681.867766] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] raise e [ 681.867766] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 681.867766] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] nwinfo = self.network_api.allocate_for_instance( [ 681.867766] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 681.867766] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] created_port_ids = self._update_ports_for_instance( [ 681.867766] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 681.867766] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] with excutils.save_and_reraise_exception(): [ 681.867766] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 681.868170] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] self.force_reraise() [ 681.868170] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 681.868170] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] raise self.value [ 681.868170] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 681.868170] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] updated_port = self._update_port( [ 681.868170] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 681.868170] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] _ensure_no_port_binding_failure(port) [ 681.868170] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 681.868170] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] raise exception.PortBindingFailed(port_id=port['id']) [ 681.868170] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] nova.exception.PortBindingFailed: Binding failed for port 24fee5ac-16fc-47a2-bba4-a3f2f983a3ca, please check neutron logs for more information. [ 681.868170] env[63202]: ERROR nova.compute.manager [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] [ 681.868550] env[63202]: DEBUG nova.compute.utils [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] Binding failed for port 24fee5ac-16fc-47a2-bba4-a3f2f983a3ca, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 681.869740] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.320s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 681.873100] env[63202]: DEBUG nova.compute.manager [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] Build of instance 45fce1e9-b1bb-48aa-9752-cbd011136abd was re-scheduled: Binding failed for port 24fee5ac-16fc-47a2-bba4-a3f2f983a3ca, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 681.873471] env[63202]: DEBUG nova.compute.manager [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 681.873700] env[63202]: DEBUG oslo_concurrency.lockutils [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "refresh_cache-45fce1e9-b1bb-48aa-9752-cbd011136abd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 681.873845] env[63202]: DEBUG oslo_concurrency.lockutils [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquired lock "refresh_cache-45fce1e9-b1bb-48aa-9752-cbd011136abd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 681.873999] env[63202]: DEBUG nova.network.neutron [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 682.043620] env[63202]: INFO nova.compute.manager [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] [instance: 4c84555a-83ac-4896-b40c-cbd7e14eebed] Took 1.03 seconds to deallocate network for instance. [ 682.076398] env[63202]: DEBUG nova.compute.manager [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 682.118924] env[63202]: DEBUG oslo_vmware.api [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385351, 'name': PowerOnVM_Task, 'duration_secs': 0.478254} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.120391] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 682.120391] env[63202]: INFO nova.compute.manager [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Took 6.32 seconds to spawn the instance on the hypervisor. [ 682.120391] env[63202]: DEBUG nova.compute.manager [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 682.120920] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37678f7c-e978-4049-a9fa-4c737c05abc6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.321379] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385352, 'name': CreateVM_Task, 'duration_secs': 0.29193} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.321563] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 682.322015] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 682.322144] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 682.322456] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 682.322693] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e70e00b9-9792-4144-9845-c6093921025a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.327442] env[63202]: DEBUG oslo_vmware.api [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Waiting for the task: (returnval){ [ 682.327442] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52640863-d4fd-2996-1d44-6b1cbf2a8c07" [ 682.327442] env[63202]: _type = "Task" [ 682.327442] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.337185] env[63202]: DEBUG oslo_vmware.api [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52640863-d4fd-2996-1d44-6b1cbf2a8c07, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.395188] env[63202]: DEBUG nova.network.neutron [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 682.537161] env[63202]: DEBUG nova.network.neutron [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.600368] env[63202]: DEBUG oslo_concurrency.lockutils [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 682.647557] env[63202]: INFO nova.compute.manager [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Took 28.35 seconds to build instance. [ 682.843746] env[63202]: DEBUG oslo_vmware.api [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52640863-d4fd-2996-1d44-6b1cbf2a8c07, 'name': SearchDatastore_Task, 'duration_secs': 0.034051} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.844942] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 682.845188] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 682.845459] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 682.845625] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 682.845825] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 682.846659] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f70893dd-fb46-4803-a150-91e4cc657080 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.849370] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e26607ae-047d-488d-8840-27d183041644 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.856190] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e23a9956-25fa-4f78-92a4-61db5088aeac {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.860237] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 682.860418] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 682.861378] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7caaa563-433f-4c72-b179-22b5468b73dd {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.889418] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4a9fb80-8c31-4f0d-9f20-2172fa80b5e6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.894163] env[63202]: DEBUG oslo_vmware.api [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Waiting for the task: (returnval){ [ 682.894163] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]525aa00c-052c-d1d5-7f95-b59f168b59c2" [ 682.894163] env[63202]: _type = "Task" [ 682.894163] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.902336] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d230b5fd-76bd-4494-8455-5736ea205ce6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.909910] env[63202]: DEBUG oslo_vmware.api [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]525aa00c-052c-d1d5-7f95-b59f168b59c2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.919408] env[63202]: DEBUG nova.compute.provider_tree [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 683.039327] env[63202]: DEBUG oslo_concurrency.lockutils [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Releasing lock "refresh_cache-45fce1e9-b1bb-48aa-9752-cbd011136abd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 683.039604] env[63202]: DEBUG nova.compute.manager [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 683.039825] env[63202]: DEBUG nova.compute.manager [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 683.040010] env[63202]: DEBUG nova.network.neutron [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 683.070473] env[63202]: DEBUG nova.network.neutron [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 683.085621] env[63202]: INFO nova.scheduler.client.report [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] Deleted allocations for instance 4c84555a-83ac-4896-b40c-cbd7e14eebed [ 683.149862] env[63202]: DEBUG oslo_concurrency.lockutils [None req-607f38d8-954a-49ca-9b71-3a858e1cbf20 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Lock "c8d27e19-0f3c-4cc3-826b-e07b3a635e5e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 64.330s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 683.407356] env[63202]: DEBUG oslo_vmware.api [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]525aa00c-052c-d1d5-7f95-b59f168b59c2, 'name': SearchDatastore_Task, 'duration_secs': 0.018063} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.408488] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f8845ab-9036-4d83-9551-3fed38a9fa75 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.415422] env[63202]: DEBUG oslo_vmware.api [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Waiting for the task: (returnval){ [ 683.415422] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]5297bc37-0d3b-e3e3-db2c-649dff0d67f0" [ 683.415422] env[63202]: _type = "Task" [ 683.415422] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.424912] env[63202]: DEBUG nova.scheduler.client.report [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 683.429594] env[63202]: DEBUG oslo_vmware.api [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]5297bc37-0d3b-e3e3-db2c-649dff0d67f0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.577953] env[63202]: DEBUG nova.network.neutron [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.598310] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d9f0c7b5-b65f-4188-913d-3da8864d7503 tempest-ServerActionsTestOtherB-1892850679 tempest-ServerActionsTestOtherB-1892850679-project-member] Lock "4c84555a-83ac-4896-b40c-cbd7e14eebed" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 71.586s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 683.653292] env[63202]: DEBUG nova.compute.manager [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 06496aad-b025-48c2-8436-03d43d1ac899] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 683.894059] env[63202]: INFO nova.compute.manager [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Rebuilding instance [ 683.928516] env[63202]: DEBUG oslo_vmware.api [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]5297bc37-0d3b-e3e3-db2c-649dff0d67f0, 'name': SearchDatastore_Task, 'duration_secs': 0.011936} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.928516] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 683.928516] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore2] e9b604fb-1d88-425c-92b7-008ddaec7c1a/e9b604fb-1d88-425c-92b7-008ddaec7c1a.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 683.928516] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6fd6aecd-72be-49e9-9f39-4748901c8b2a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.931045] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.061s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 683.932396] env[63202]: ERROR nova.compute.manager [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6fc2c5e8-8533-4f55-9c73-58ec49bc25fa, please check neutron logs for more information. [ 683.932396] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] Traceback (most recent call last): [ 683.932396] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 683.932396] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] self.driver.spawn(context, instance, image_meta, [ 683.932396] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 683.932396] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] self._vmops.spawn(context, instance, image_meta, injected_files, [ 683.932396] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 683.932396] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] vm_ref = self.build_virtual_machine(instance, [ 683.932396] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 683.932396] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] vif_infos = vmwarevif.get_vif_info(self._session, [ 683.932396] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 683.932834] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] for vif in network_info: [ 683.932834] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 683.932834] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] return self._sync_wrapper(fn, *args, **kwargs) [ 683.932834] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 683.932834] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] self.wait() [ 683.932834] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 683.932834] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] self[:] = self._gt.wait() [ 683.932834] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 683.932834] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] return self._exit_event.wait() [ 683.932834] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 683.932834] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] current.throw(*self._exc) [ 683.932834] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 683.932834] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] result = function(*args, **kwargs) [ 683.933168] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 683.933168] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] return func(*args, **kwargs) [ 683.933168] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 683.933168] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] raise e [ 683.933168] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 683.933168] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] nwinfo = self.network_api.allocate_for_instance( [ 683.933168] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 683.933168] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] created_port_ids = self._update_ports_for_instance( [ 683.933168] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 683.933168] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] with excutils.save_and_reraise_exception(): [ 683.933168] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 683.933168] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] self.force_reraise() [ 683.933168] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 683.933489] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] raise self.value [ 683.933489] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 683.933489] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] updated_port = self._update_port( [ 683.933489] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 683.933489] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] _ensure_no_port_binding_failure(port) [ 683.933489] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 683.933489] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] raise exception.PortBindingFailed(port_id=port['id']) [ 683.933489] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] nova.exception.PortBindingFailed: Binding failed for port 6fc2c5e8-8533-4f55-9c73-58ec49bc25fa, please check neutron logs for more information. [ 683.933489] env[63202]: ERROR nova.compute.manager [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] [ 683.933489] env[63202]: DEBUG nova.compute.utils [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] Binding failed for port 6fc2c5e8-8533-4f55-9c73-58ec49bc25fa, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 683.937191] env[63202]: DEBUG oslo_concurrency.lockutils [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.781s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 683.938718] env[63202]: INFO nova.compute.claims [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 683.943355] env[63202]: DEBUG nova.compute.manager [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] Build of instance ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93 was re-scheduled: Binding failed for port 6fc2c5e8-8533-4f55-9c73-58ec49bc25fa, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 683.943783] env[63202]: DEBUG nova.compute.manager [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 683.944010] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] Acquiring lock "refresh_cache-ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 683.944183] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] Acquired lock "refresh_cache-ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 683.944340] env[63202]: DEBUG nova.network.neutron [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 683.947561] env[63202]: DEBUG oslo_vmware.api [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Waiting for the task: (returnval){ [ 683.947561] env[63202]: value = "task-1385353" [ 683.947561] env[63202]: _type = "Task" [ 683.947561] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.956562] env[63202]: DEBUG oslo_vmware.api [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Task: {'id': task-1385353, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.966674] env[63202]: DEBUG nova.compute.manager [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 683.967657] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f45e605-f028-45c1-9da9-131f7d71fda3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.080719] env[63202]: INFO nova.compute.manager [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 45fce1e9-b1bb-48aa-9752-cbd011136abd] Took 1.04 seconds to deallocate network for instance. [ 684.101190] env[63202]: DEBUG nova.compute.manager [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 684.177480] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 684.460321] env[63202]: DEBUG oslo_vmware.api [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Task: {'id': task-1385353, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.49739} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.460620] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore2] e9b604fb-1d88-425c-92b7-008ddaec7c1a/e9b604fb-1d88-425c-92b7-008ddaec7c1a.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 684.460896] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 684.461161] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b4402613-b5d6-4636-b655-f361ee9debc1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.468406] env[63202]: DEBUG oslo_vmware.api [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Waiting for the task: (returnval){ [ 684.468406] env[63202]: value = "task-1385354" [ 684.468406] env[63202]: _type = "Task" [ 684.468406] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.478815] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 684.479271] env[63202]: DEBUG oslo_vmware.api [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Task: {'id': task-1385354, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.479631] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a6fc6d93-c093-4e3e-a131-4df2791f634a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.482319] env[63202]: DEBUG nova.network.neutron [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 684.489353] env[63202]: DEBUG oslo_vmware.api [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Waiting for the task: (returnval){ [ 684.489353] env[63202]: value = "task-1385355" [ 684.489353] env[63202]: _type = "Task" [ 684.489353] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.498959] env[63202]: DEBUG oslo_vmware.api [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385355, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.626264] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 684.669045] env[63202]: DEBUG nova.network.neutron [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.981364] env[63202]: DEBUG oslo_vmware.api [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Task: {'id': task-1385354, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060275} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.981648] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 684.982580] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6e64b89-8663-414b-8fd7-ea29569d67d4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.006725] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Reconfiguring VM instance instance-0000001a to attach disk [datastore2] e9b604fb-1d88-425c-92b7-008ddaec7c1a/e9b604fb-1d88-425c-92b7-008ddaec7c1a.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 685.013372] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8eda3573-448b-416b-964f-6bee7712cc82 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.033723] env[63202]: DEBUG oslo_vmware.api [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385355, 'name': PowerOffVM_Task, 'duration_secs': 0.110912} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.034217] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 685.034430] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 685.034737] env[63202]: DEBUG oslo_vmware.api [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Waiting for the task: (returnval){ [ 685.034737] env[63202]: value = "task-1385356" [ 685.034737] env[63202]: _type = "Task" [ 685.034737] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.035474] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a7faeae-22cb-452b-931f-8e251a44bd3f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.050246] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 685.050542] env[63202]: DEBUG oslo_vmware.api [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Task: {'id': task-1385356, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.050799] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-82c04686-8efd-4590-859f-a418f28a8cf5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.076092] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 685.076404] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Deleting contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 685.076629] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Deleting the datastore file [datastore1] c8d27e19-0f3c-4cc3-826b-e07b3a635e5e {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 685.076853] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4faefcc0-b691-4eaf-a794-256669b8d55c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.085750] env[63202]: DEBUG oslo_vmware.api [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Waiting for the task: (returnval){ [ 685.085750] env[63202]: value = "task-1385358" [ 685.085750] env[63202]: _type = "Task" [ 685.085750] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.098879] env[63202]: DEBUG oslo_vmware.api [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385358, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.134642] env[63202]: INFO nova.scheduler.client.report [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Deleted allocations for instance 45fce1e9-b1bb-48aa-9752-cbd011136abd [ 685.171915] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] Releasing lock "refresh_cache-ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 685.171915] env[63202]: DEBUG nova.compute.manager [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 685.171915] env[63202]: DEBUG nova.compute.manager [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 685.171915] env[63202]: DEBUG nova.network.neutron [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 685.194416] env[63202]: DEBUG nova.network.neutron [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 685.427490] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a56a67c1-2ddc-4f18-bd47-c7ae507c29a6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.435182] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eff5128-5bde-4614-b161-c3ddf38ac963 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.467687] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce517e68-6788-46ba-99ae-c1745f607a61 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.474765] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff20bc3c-7ee9-4f1c-a5cd-9a32d2dbc0ea {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.489846] env[63202]: DEBUG nova.compute.provider_tree [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 685.548166] env[63202]: DEBUG oslo_vmware.api [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Task: {'id': task-1385356, 'name': ReconfigVM_Task, 'duration_secs': 0.277949} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.548166] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Reconfigured VM instance instance-0000001a to attach disk [datastore2] e9b604fb-1d88-425c-92b7-008ddaec7c1a/e9b604fb-1d88-425c-92b7-008ddaec7c1a.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 685.548166] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ac7889af-6870-44c6-869e-d042d7380ca9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.554515] env[63202]: DEBUG oslo_vmware.api [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Waiting for the task: (returnval){ [ 685.554515] env[63202]: value = "task-1385359" [ 685.554515] env[63202]: _type = "Task" [ 685.554515] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.563602] env[63202]: DEBUG oslo_vmware.api [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Task: {'id': task-1385359, 'name': Rename_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.597776] env[63202]: DEBUG oslo_vmware.api [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385358, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.102709} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.597776] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 685.597776] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Deleted contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 685.597776] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 685.649429] env[63202]: DEBUG oslo_concurrency.lockutils [None req-93b9d231-933b-47b0-95d3-4d562b2c9cd6 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "45fce1e9-b1bb-48aa-9752-cbd011136abd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 71.227s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 685.697396] env[63202]: DEBUG nova.network.neutron [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 685.812553] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Acquiring lock "06bd8147-5479-4c5e-8ba2-6d3a205ba05e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 685.812780] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Lock "06bd8147-5479-4c5e-8ba2-6d3a205ba05e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 685.993214] env[63202]: DEBUG nova.scheduler.client.report [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 686.067348] env[63202]: DEBUG oslo_vmware.api [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Task: {'id': task-1385359, 'name': Rename_Task, 'duration_secs': 0.126353} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 686.067648] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 686.067973] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-761b0636-e587-4597-b79e-1b54996e8815 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.074618] env[63202]: DEBUG oslo_vmware.api [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Waiting for the task: (returnval){ [ 686.074618] env[63202]: value = "task-1385360" [ 686.074618] env[63202]: _type = "Task" [ 686.074618] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 686.084279] env[63202]: DEBUG oslo_vmware.api [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Task: {'id': task-1385360, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.156717] env[63202]: DEBUG nova.compute.manager [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] [instance: a97cff65-e350-480e-9891-3317fe05be47] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 686.165867] env[63202]: DEBUG oslo_concurrency.lockutils [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] Acquiring lock "56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 686.166280] env[63202]: DEBUG oslo_concurrency.lockutils [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] Lock "56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 686.203457] env[63202]: INFO nova.compute.manager [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] [instance: ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93] Took 1.03 seconds to deallocate network for instance. [ 686.501335] env[63202]: DEBUG oslo_concurrency.lockutils [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.563s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 686.501335] env[63202]: DEBUG nova.compute.manager [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 686.504009] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 17.101s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 686.587866] env[63202]: DEBUG oslo_vmware.api [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Task: {'id': task-1385360, 'name': PowerOnVM_Task, 'duration_secs': 0.445755} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 686.588790] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 686.589183] env[63202]: DEBUG nova.compute.manager [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 686.590050] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e502c13d-98f3-45a8-b607-9ddead38a003 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.640199] env[63202]: DEBUG nova.virt.hardware [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 686.640432] env[63202]: DEBUG nova.virt.hardware [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 686.640576] env[63202]: DEBUG nova.virt.hardware [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 686.640765] env[63202]: DEBUG nova.virt.hardware [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 686.640901] env[63202]: DEBUG nova.virt.hardware [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 686.641126] env[63202]: DEBUG nova.virt.hardware [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 686.641306] env[63202]: DEBUG nova.virt.hardware [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 686.641458] env[63202]: DEBUG nova.virt.hardware [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 686.641688] env[63202]: DEBUG nova.virt.hardware [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 686.641764] env[63202]: DEBUG nova.virt.hardware [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 686.641923] env[63202]: DEBUG nova.virt.hardware [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 686.642974] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-518d99e6-342f-484b-b66b-00c77cb30eb9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.650699] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-699862ba-545c-4f99-a865-79949529c89a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.667213] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Instance VIF info [] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 686.672867] env[63202]: DEBUG oslo.service.loopingcall [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 686.676238] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 686.676238] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-aa2b67a4-828a-4535-98c2-9940b4959108 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.688736] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 686.694495] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 686.694495] env[63202]: value = "task-1385361" [ 686.694495] env[63202]: _type = "Task" [ 686.694495] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 686.702134] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385361, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.008225] env[63202]: DEBUG nova.compute.utils [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 687.009332] env[63202]: DEBUG nova.compute.manager [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 687.010091] env[63202]: DEBUG nova.network.neutron [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 687.049683] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.050066] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.079909] env[63202]: DEBUG nova.policy [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '46863ba14f1c4829a24eca823e763f2f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b028e34225744668807aae95712a8d41', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 687.108855] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.204546] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385361, 'name': CreateVM_Task} progress is 99%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.236996] env[63202]: INFO nova.scheduler.client.report [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] Deleted allocations for instance ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93 [ 687.439787] env[63202]: DEBUG nova.network.neutron [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] Successfully created port: 5d0c4734-7baa-410e-9f3e-91d875f0248b {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 687.512684] env[63202]: DEBUG nova.compute.manager [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 687.535731] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 8cd0415d-dab9-4554-966c-1ea8d822deff actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 687.535879] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance e9b604fb-1d88-425c-92b7-008ddaec7c1a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 687.535999] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance c8d27e19-0f3c-4cc3-826b-e07b3a635e5e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 687.536127] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 687.707021] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385361, 'name': CreateVM_Task} progress is 99%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.748404] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4f754c03-2736-4497-b917-75ddc7148c93 tempest-ServerActionsTestJSON-302989054 tempest-ServerActionsTestJSON-302989054-project-member] Lock "ba17f6c3-8a5e-4be8-b6cb-ad0234cdbe93" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 688.039119] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 864e1f52-f77e-4596-a5b5-dff3679f911f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 688.206329] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385361, 'name': CreateVM_Task, 'duration_secs': 1.283759} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.206456] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 688.206876] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 688.207103] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.207430] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 688.207675] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ffe87fa6-983b-4814-a075-f20e66eb6b93 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.213972] env[63202]: DEBUG oslo_vmware.api [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Waiting for the task: (returnval){ [ 688.213972] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]5260fd20-0b51-6c79-cf09-4dd7d3b773f1" [ 688.213972] env[63202]: _type = "Task" [ 688.213972] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 688.221740] env[63202]: DEBUG oslo_vmware.api [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]5260fd20-0b51-6c79-cf09-4dd7d3b773f1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.250372] env[63202]: DEBUG nova.compute.manager [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 688.377692] env[63202]: DEBUG oslo_concurrency.lockutils [None req-9c0370f1-864c-44a4-bb17-94eac1e1a874 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Acquiring lock "e9b604fb-1d88-425c-92b7-008ddaec7c1a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 688.377967] env[63202]: DEBUG oslo_concurrency.lockutils [None req-9c0370f1-864c-44a4-bb17-94eac1e1a874 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Lock "e9b604fb-1d88-425c-92b7-008ddaec7c1a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 688.378255] env[63202]: DEBUG oslo_concurrency.lockutils [None req-9c0370f1-864c-44a4-bb17-94eac1e1a874 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Acquiring lock "e9b604fb-1d88-425c-92b7-008ddaec7c1a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 688.378446] env[63202]: DEBUG oslo_concurrency.lockutils [None req-9c0370f1-864c-44a4-bb17-94eac1e1a874 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Lock "e9b604fb-1d88-425c-92b7-008ddaec7c1a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 688.378658] env[63202]: DEBUG oslo_concurrency.lockutils [None req-9c0370f1-864c-44a4-bb17-94eac1e1a874 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Lock "e9b604fb-1d88-425c-92b7-008ddaec7c1a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 688.384394] env[63202]: INFO nova.compute.manager [None req-9c0370f1-864c-44a4-bb17-94eac1e1a874 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Terminating instance [ 688.386341] env[63202]: DEBUG oslo_concurrency.lockutils [None req-9c0370f1-864c-44a4-bb17-94eac1e1a874 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Acquiring lock "refresh_cache-e9b604fb-1d88-425c-92b7-008ddaec7c1a" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 688.386518] env[63202]: DEBUG oslo_concurrency.lockutils [None req-9c0370f1-864c-44a4-bb17-94eac1e1a874 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Acquired lock "refresh_cache-e9b604fb-1d88-425c-92b7-008ddaec7c1a" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.386665] env[63202]: DEBUG nova.network.neutron [None req-9c0370f1-864c-44a4-bb17-94eac1e1a874 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 688.460212] env[63202]: DEBUG nova.compute.manager [req-a1bf2f42-4b65-4add-9ab3-81019e0feba7 req-a3103a49-1f58-4748-a47e-2fdf3f229871 service nova] [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] Received event network-changed-5d0c4734-7baa-410e-9f3e-91d875f0248b {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 688.460405] env[63202]: DEBUG nova.compute.manager [req-a1bf2f42-4b65-4add-9ab3-81019e0feba7 req-a3103a49-1f58-4748-a47e-2fdf3f229871 service nova] [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] Refreshing instance network info cache due to event network-changed-5d0c4734-7baa-410e-9f3e-91d875f0248b. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 688.460606] env[63202]: DEBUG oslo_concurrency.lockutils [req-a1bf2f42-4b65-4add-9ab3-81019e0feba7 req-a3103a49-1f58-4748-a47e-2fdf3f229871 service nova] Acquiring lock "refresh_cache-ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 688.460888] env[63202]: DEBUG oslo_concurrency.lockutils [req-a1bf2f42-4b65-4add-9ab3-81019e0feba7 req-a3103a49-1f58-4748-a47e-2fdf3f229871 service nova] Acquired lock "refresh_cache-ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.462301] env[63202]: DEBUG nova.network.neutron [req-a1bf2f42-4b65-4add-9ab3-81019e0feba7 req-a3103a49-1f58-4748-a47e-2fdf3f229871 service nova] [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] Refreshing network info cache for port 5d0c4734-7baa-410e-9f3e-91d875f0248b {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 688.526795] env[63202]: DEBUG nova.compute.manager [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 688.545512] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance eff8282a-b3cd-4038-9f48-94cd4c4afc55 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 688.559791] env[63202]: DEBUG nova.virt.hardware [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 688.560101] env[63202]: DEBUG nova.virt.hardware [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 688.560282] env[63202]: DEBUG nova.virt.hardware [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 688.560464] env[63202]: DEBUG nova.virt.hardware [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 688.561572] env[63202]: DEBUG nova.virt.hardware [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 688.561572] env[63202]: DEBUG nova.virt.hardware [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 688.561572] env[63202]: DEBUG nova.virt.hardware [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 688.561572] env[63202]: DEBUG nova.virt.hardware [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 688.561572] env[63202]: DEBUG nova.virt.hardware [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 688.561788] env[63202]: DEBUG nova.virt.hardware [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 688.561788] env[63202]: DEBUG nova.virt.hardware [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 688.562551] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b215c07-9ffb-4775-b452-6b22cbb91710 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.573258] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07c68e88-d5be-4a0e-8c36-338760b898bd {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.635700] env[63202]: ERROR nova.compute.manager [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5d0c4734-7baa-410e-9f3e-91d875f0248b, please check neutron logs for more information. [ 688.635700] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 688.635700] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 688.635700] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 688.635700] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 688.635700] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 688.635700] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 688.635700] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 688.635700] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 688.635700] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 688.635700] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 688.635700] env[63202]: ERROR nova.compute.manager raise self.value [ 688.635700] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 688.635700] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 688.635700] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 688.635700] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 688.636220] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 688.636220] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 688.636220] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5d0c4734-7baa-410e-9f3e-91d875f0248b, please check neutron logs for more information. [ 688.636220] env[63202]: ERROR nova.compute.manager [ 688.636220] env[63202]: Traceback (most recent call last): [ 688.636220] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 688.636220] env[63202]: listener.cb(fileno) [ 688.636220] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 688.636220] env[63202]: result = function(*args, **kwargs) [ 688.636220] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 688.636220] env[63202]: return func(*args, **kwargs) [ 688.636220] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 688.636220] env[63202]: raise e [ 688.636220] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 688.636220] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 688.636220] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 688.636220] env[63202]: created_port_ids = self._update_ports_for_instance( [ 688.636220] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 688.636220] env[63202]: with excutils.save_and_reraise_exception(): [ 688.636220] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 688.636220] env[63202]: self.force_reraise() [ 688.636220] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 688.636220] env[63202]: raise self.value [ 688.636220] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 688.636220] env[63202]: updated_port = self._update_port( [ 688.636220] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 688.636220] env[63202]: _ensure_no_port_binding_failure(port) [ 688.636220] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 688.636220] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 688.637088] env[63202]: nova.exception.PortBindingFailed: Binding failed for port 5d0c4734-7baa-410e-9f3e-91d875f0248b, please check neutron logs for more information. [ 688.637088] env[63202]: Removing descriptor: 16 [ 688.637088] env[63202]: ERROR nova.compute.manager [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5d0c4734-7baa-410e-9f3e-91d875f0248b, please check neutron logs for more information. [ 688.637088] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] Traceback (most recent call last): [ 688.637088] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 688.637088] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] yield resources [ 688.637088] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 688.637088] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] self.driver.spawn(context, instance, image_meta, [ 688.637088] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 688.637088] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] self._vmops.spawn(context, instance, image_meta, injected_files, [ 688.637088] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 688.637088] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] vm_ref = self.build_virtual_machine(instance, [ 688.637418] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 688.637418] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] vif_infos = vmwarevif.get_vif_info(self._session, [ 688.637418] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 688.637418] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] for vif in network_info: [ 688.637418] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 688.637418] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] return self._sync_wrapper(fn, *args, **kwargs) [ 688.637418] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 688.637418] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] self.wait() [ 688.637418] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 688.637418] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] self[:] = self._gt.wait() [ 688.637418] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 688.637418] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] return self._exit_event.wait() [ 688.637418] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 688.637780] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] result = hub.switch() [ 688.637780] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 688.637780] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] return self.greenlet.switch() [ 688.637780] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 688.637780] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] result = function(*args, **kwargs) [ 688.637780] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 688.637780] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] return func(*args, **kwargs) [ 688.637780] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 688.637780] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] raise e [ 688.637780] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 688.637780] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] nwinfo = self.network_api.allocate_for_instance( [ 688.637780] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 688.637780] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] created_port_ids = self._update_ports_for_instance( [ 688.638171] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 688.638171] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] with excutils.save_and_reraise_exception(): [ 688.638171] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 688.638171] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] self.force_reraise() [ 688.638171] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 688.638171] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] raise self.value [ 688.638171] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 688.638171] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] updated_port = self._update_port( [ 688.638171] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 688.638171] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] _ensure_no_port_binding_failure(port) [ 688.638171] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 688.638171] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] raise exception.PortBindingFailed(port_id=port['id']) [ 688.638497] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] nova.exception.PortBindingFailed: Binding failed for port 5d0c4734-7baa-410e-9f3e-91d875f0248b, please check neutron logs for more information. [ 688.638497] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] [ 688.638497] env[63202]: INFO nova.compute.manager [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] Terminating instance [ 688.638877] env[63202]: DEBUG oslo_concurrency.lockutils [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Acquiring lock "refresh_cache-ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 688.725768] env[63202]: DEBUG oslo_vmware.api [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]5260fd20-0b51-6c79-cf09-4dd7d3b773f1, 'name': SearchDatastore_Task, 'duration_secs': 0.009558} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.726104] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 688.726296] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 688.726526] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 688.726663] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.726892] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 688.727089] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-876d21dd-aef1-4ae8-ab2f-edb8d31b303b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.735017] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 688.735194] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 688.735957] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49ffc113-b7a2-433e-ab79-7b368a8e2d02 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.740891] env[63202]: DEBUG oslo_vmware.api [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Waiting for the task: (returnval){ [ 688.740891] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]526830df-b908-539b-57bc-4fa548658f18" [ 688.740891] env[63202]: _type = "Task" [ 688.740891] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 688.748536] env[63202]: DEBUG oslo_vmware.api [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]526830df-b908-539b-57bc-4fa548658f18, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.775310] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 688.905607] env[63202]: DEBUG nova.network.neutron [None req-9c0370f1-864c-44a4-bb17-94eac1e1a874 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 688.966219] env[63202]: DEBUG nova.network.neutron [None req-9c0370f1-864c-44a4-bb17-94eac1e1a874 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.985451] env[63202]: DEBUG nova.network.neutron [req-a1bf2f42-4b65-4add-9ab3-81019e0feba7 req-a3103a49-1f58-4748-a47e-2fdf3f229871 service nova] [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 689.049021] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance de5f9555-3959-41cd-b63a-db128a001631 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 689.143918] env[63202]: DEBUG nova.network.neutron [req-a1bf2f42-4b65-4add-9ab3-81019e0feba7 req-a3103a49-1f58-4748-a47e-2fdf3f229871 service nova] [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.251244] env[63202]: DEBUG oslo_vmware.api [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]526830df-b908-539b-57bc-4fa548658f18, 'name': SearchDatastore_Task, 'duration_secs': 0.008459} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.252127] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f525bc47-f265-463a-995e-c00ec66e6024 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.257463] env[63202]: DEBUG oslo_vmware.api [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Waiting for the task: (returnval){ [ 689.257463] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52d0b1c6-4534-95a0-fcf7-f1521fb3b9e2" [ 689.257463] env[63202]: _type = "Task" [ 689.257463] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.265235] env[63202]: DEBUG oslo_vmware.api [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52d0b1c6-4534-95a0-fcf7-f1521fb3b9e2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.468946] env[63202]: DEBUG oslo_concurrency.lockutils [None req-9c0370f1-864c-44a4-bb17-94eac1e1a874 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Releasing lock "refresh_cache-e9b604fb-1d88-425c-92b7-008ddaec7c1a" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 689.469349] env[63202]: DEBUG nova.compute.manager [None req-9c0370f1-864c-44a4-bb17-94eac1e1a874 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 689.469500] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-9c0370f1-864c-44a4-bb17-94eac1e1a874 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 689.470387] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b78bb519-72a6-4724-8d0d-ed92cedc0330 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.477876] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c0370f1-864c-44a4-bb17-94eac1e1a874 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 689.478169] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-11e7821b-5bab-4e04-9d4a-e38c00973c44 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.485158] env[63202]: DEBUG oslo_vmware.api [None req-9c0370f1-864c-44a4-bb17-94eac1e1a874 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Waiting for the task: (returnval){ [ 689.485158] env[63202]: value = "task-1385362" [ 689.485158] env[63202]: _type = "Task" [ 689.485158] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.493943] env[63202]: DEBUG oslo_vmware.api [None req-9c0370f1-864c-44a4-bb17-94eac1e1a874 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Task: {'id': task-1385362, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.553647] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 2f87939c-d615-4d13-8695-971e83ca7843 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 689.646944] env[63202]: DEBUG oslo_concurrency.lockutils [req-a1bf2f42-4b65-4add-9ab3-81019e0feba7 req-a3103a49-1f58-4748-a47e-2fdf3f229871 service nova] Releasing lock "refresh_cache-ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 689.647229] env[63202]: DEBUG oslo_concurrency.lockutils [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Acquired lock "refresh_cache-ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 689.647406] env[63202]: DEBUG nova.network.neutron [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 689.773021] env[63202]: DEBUG oslo_vmware.api [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52d0b1c6-4534-95a0-fcf7-f1521fb3b9e2, 'name': SearchDatastore_Task, 'duration_secs': 0.009017} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.773021] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 689.773021] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] c8d27e19-0f3c-4cc3-826b-e07b3a635e5e/c8d27e19-0f3c-4cc3-826b-e07b3a635e5e.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 689.773021] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b9659ba1-a873-4111-a211-398beb730669 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.778399] env[63202]: DEBUG oslo_vmware.api [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Waiting for the task: (returnval){ [ 689.778399] env[63202]: value = "task-1385363" [ 689.778399] env[63202]: _type = "Task" [ 689.778399] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.788778] env[63202]: DEBUG oslo_vmware.api [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385363, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.995394] env[63202]: DEBUG oslo_vmware.api [None req-9c0370f1-864c-44a4-bb17-94eac1e1a874 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Task: {'id': task-1385362, 'name': PowerOffVM_Task, 'duration_secs': 0.184661} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.996793] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c0370f1-864c-44a4-bb17-94eac1e1a874 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 689.996793] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-9c0370f1-864c-44a4-bb17-94eac1e1a874 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 689.996793] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9c83ca99-4db9-45b6-b265-9d01fae8f517 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.025916] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-9c0370f1-864c-44a4-bb17-94eac1e1a874 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 690.025916] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-9c0370f1-864c-44a4-bb17-94eac1e1a874 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Deleting contents of the VM from datastore datastore2 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 690.026319] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c0370f1-864c-44a4-bb17-94eac1e1a874 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Deleting the datastore file [datastore2] e9b604fb-1d88-425c-92b7-008ddaec7c1a {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 690.026657] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2e607590-f214-439a-accd-d879d29ac452 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.033593] env[63202]: DEBUG oslo_vmware.api [None req-9c0370f1-864c-44a4-bb17-94eac1e1a874 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Waiting for the task: (returnval){ [ 690.033593] env[63202]: value = "task-1385365" [ 690.033593] env[63202]: _type = "Task" [ 690.033593] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.042340] env[63202]: DEBUG oslo_vmware.api [None req-9c0370f1-864c-44a4-bb17-94eac1e1a874 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Task: {'id': task-1385365, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.061560] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 690.174496] env[63202]: DEBUG nova.network.neutron [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 690.289085] env[63202]: DEBUG oslo_vmware.api [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385363, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.460379} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.289355] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] c8d27e19-0f3c-4cc3-826b-e07b3a635e5e/c8d27e19-0f3c-4cc3-826b-e07b3a635e5e.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 690.289558] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 690.289793] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7a30217f-99e3-4739-9087-9939c990da1d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.295454] env[63202]: DEBUG oslo_vmware.api [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Waiting for the task: (returnval){ [ 690.295454] env[63202]: value = "task-1385366" [ 690.295454] env[63202]: _type = "Task" [ 690.295454] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.303753] env[63202]: DEBUG nova.network.neutron [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.309826] env[63202]: DEBUG oslo_vmware.api [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385366, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.484584] env[63202]: DEBUG nova.compute.manager [req-d99abf2a-ddeb-4208-9176-098425fe5269 req-1ee446ec-54d2-4cbd-8c66-103de5688af5 service nova] [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] Received event network-vif-deleted-5d0c4734-7baa-410e-9f3e-91d875f0248b {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 690.542756] env[63202]: DEBUG oslo_vmware.api [None req-9c0370f1-864c-44a4-bb17-94eac1e1a874 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Task: {'id': task-1385365, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.227012} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.543014] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c0370f1-864c-44a4-bb17-94eac1e1a874 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 690.543211] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-9c0370f1-864c-44a4-bb17-94eac1e1a874 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Deleted contents of the VM from datastore datastore2 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 690.543390] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-9c0370f1-864c-44a4-bb17-94eac1e1a874 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 690.543556] env[63202]: INFO nova.compute.manager [None req-9c0370f1-864c-44a4-bb17-94eac1e1a874 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Took 1.07 seconds to destroy the instance on the hypervisor. [ 690.543782] env[63202]: DEBUG oslo.service.loopingcall [None req-9c0370f1-864c-44a4-bb17-94eac1e1a874 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 690.543965] env[63202]: DEBUG nova.compute.manager [-] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 690.544069] env[63202]: DEBUG nova.network.neutron [-] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 690.562879] env[63202]: DEBUG nova.network.neutron [-] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 690.571512] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 06496aad-b025-48c2-8436-03d43d1ac899 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 690.804936] env[63202]: DEBUG oslo_vmware.api [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385366, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059321} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.805228] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 690.805985] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49df4089-74d8-42e5-8f77-d57a87ad96a9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.816905] env[63202]: DEBUG oslo_concurrency.lockutils [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Releasing lock "refresh_cache-ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 690.817385] env[63202]: DEBUG nova.compute.manager [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 690.817627] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 690.827098] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Reconfiguring VM instance instance-0000001b to attach disk [datastore1] c8d27e19-0f3c-4cc3-826b-e07b3a635e5e/c8d27e19-0f3c-4cc3-826b-e07b3a635e5e.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 690.827098] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3d3ac0a6-106a-40c3-adb4-02992786bded {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.829025] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ca3300b6-c030-499c-92c8-9000f21cc51e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.849951] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a74b6c0-0bb5-471e-8b4b-80ea478bec79 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.860982] env[63202]: DEBUG oslo_vmware.api [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Waiting for the task: (returnval){ [ 690.860982] env[63202]: value = "task-1385367" [ 690.860982] env[63202]: _type = "Task" [ 690.860982] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.869187] env[63202]: DEBUG oslo_vmware.api [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385367, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.874840] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88 could not be found. [ 690.875078] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 690.875197] env[63202]: INFO nova.compute.manager [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] Took 0.06 seconds to destroy the instance on the hypervisor. [ 690.875442] env[63202]: DEBUG oslo.service.loopingcall [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 690.875671] env[63202]: DEBUG nova.compute.manager [-] [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 690.875754] env[63202]: DEBUG nova.network.neutron [-] [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 690.891477] env[63202]: DEBUG nova.network.neutron [-] [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 691.067291] env[63202]: DEBUG nova.network.neutron [-] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.074120] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 8885156e-78f1-45ff-87c9-829f3ab89dca has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 691.371753] env[63202]: DEBUG oslo_vmware.api [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385367, 'name': ReconfigVM_Task, 'duration_secs': 0.318935} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.371953] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Reconfigured VM instance instance-0000001b to attach disk [datastore1] c8d27e19-0f3c-4cc3-826b-e07b3a635e5e/c8d27e19-0f3c-4cc3-826b-e07b3a635e5e.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 691.372524] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-55fa4ac7-7709-481e-a146-7edd57275107 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.379914] env[63202]: DEBUG oslo_vmware.api [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Waiting for the task: (returnval){ [ 691.379914] env[63202]: value = "task-1385368" [ 691.379914] env[63202]: _type = "Task" [ 691.379914] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.390013] env[63202]: DEBUG oslo_vmware.api [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385368, 'name': Rename_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.395038] env[63202]: DEBUG nova.network.neutron [-] [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.568684] env[63202]: INFO nova.compute.manager [-] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Took 1.02 seconds to deallocate network for instance. [ 691.576981] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance a97cff65-e350-480e-9891-3317fe05be47 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 691.890209] env[63202]: DEBUG oslo_vmware.api [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385368, 'name': Rename_Task, 'duration_secs': 0.137247} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.890482] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 691.890715] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-33762f22-ca78-4af4-9529-829cb395ed56 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.896798] env[63202]: INFO nova.compute.manager [-] [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] Took 1.02 seconds to deallocate network for instance. [ 691.899444] env[63202]: DEBUG oslo_vmware.api [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Waiting for the task: (returnval){ [ 691.899444] env[63202]: value = "task-1385369" [ 691.899444] env[63202]: _type = "Task" [ 691.899444] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.899868] env[63202]: DEBUG nova.compute.claims [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 691.900036] env[63202]: DEBUG oslo_concurrency.lockutils [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 691.907613] env[63202]: DEBUG oslo_vmware.api [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385369, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.076156] env[63202]: DEBUG oslo_concurrency.lockutils [None req-9c0370f1-864c-44a4-bb17-94eac1e1a874 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 692.080102] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 692.417477] env[63202]: DEBUG oslo_vmware.api [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385369, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.583492] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 611fead3-3d1b-41e4-9579-7ad1a1b754c3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 692.910768] env[63202]: DEBUG oslo_vmware.api [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385369, 'name': PowerOnVM_Task, 'duration_secs': 0.617994} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.911115] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 692.911115] env[63202]: DEBUG nova.compute.manager [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 692.911910] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-295fb75b-3ae5-4332-b4c2-4a576f0fad30 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.086142] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 4cd7f7b3-d947-4745-8fd7-940076865e3b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 693.435800] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 693.591931] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance ba2d047d-d30f-4ba0-bcfb-787c5a3ae516 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 694.096463] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 1ac8e024-65d1-4250-a946-858aaff852c8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 694.319061] env[63202]: DEBUG oslo_concurrency.lockutils [None req-31c518cb-d896-429a-88e7-bb67df082013 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Acquiring lock "c8d27e19-0f3c-4cc3-826b-e07b3a635e5e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 694.319431] env[63202]: DEBUG oslo_concurrency.lockutils [None req-31c518cb-d896-429a-88e7-bb67df082013 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Lock "c8d27e19-0f3c-4cc3-826b-e07b3a635e5e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 694.319648] env[63202]: DEBUG oslo_concurrency.lockutils [None req-31c518cb-d896-429a-88e7-bb67df082013 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Acquiring lock "c8d27e19-0f3c-4cc3-826b-e07b3a635e5e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 694.319826] env[63202]: DEBUG oslo_concurrency.lockutils [None req-31c518cb-d896-429a-88e7-bb67df082013 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Lock "c8d27e19-0f3c-4cc3-826b-e07b3a635e5e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 694.320016] env[63202]: DEBUG oslo_concurrency.lockutils [None req-31c518cb-d896-429a-88e7-bb67df082013 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Lock "c8d27e19-0f3c-4cc3-826b-e07b3a635e5e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 694.325371] env[63202]: INFO nova.compute.manager [None req-31c518cb-d896-429a-88e7-bb67df082013 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Terminating instance [ 694.327222] env[63202]: DEBUG oslo_concurrency.lockutils [None req-31c518cb-d896-429a-88e7-bb67df082013 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Acquiring lock "refresh_cache-c8d27e19-0f3c-4cc3-826b-e07b3a635e5e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 694.327374] env[63202]: DEBUG oslo_concurrency.lockutils [None req-31c518cb-d896-429a-88e7-bb67df082013 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Acquired lock "refresh_cache-c8d27e19-0f3c-4cc3-826b-e07b3a635e5e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 694.327698] env[63202]: DEBUG nova.network.neutron [None req-31c518cb-d896-429a-88e7-bb67df082013 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 694.599334] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 3080314c-938e-4c27-bffd-547bdc7e6e38 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 694.850913] env[63202]: DEBUG nova.network.neutron [None req-31c518cb-d896-429a-88e7-bb67df082013 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 694.911666] env[63202]: DEBUG nova.network.neutron [None req-31c518cb-d896-429a-88e7-bb67df082013 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.982187] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] Acquiring lock "a47dbb3e-62be-426d-acd7-bedc0faf65c5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 694.982516] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] Lock "a47dbb3e-62be-426d-acd7-bedc0faf65c5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 695.102634] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 7efc454e-337d-43db-9076-bfc5b89eeea4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 695.416039] env[63202]: DEBUG oslo_concurrency.lockutils [None req-31c518cb-d896-429a-88e7-bb67df082013 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Releasing lock "refresh_cache-c8d27e19-0f3c-4cc3-826b-e07b3a635e5e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 695.416358] env[63202]: DEBUG nova.compute.manager [None req-31c518cb-d896-429a-88e7-bb67df082013 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 695.416553] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-31c518cb-d896-429a-88e7-bb67df082013 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 695.417464] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d284ffae-76d6-4828-92b2-9dddc3744104 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.425014] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-31c518cb-d896-429a-88e7-bb67df082013 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 695.425262] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dbaf01c3-f03c-48f5-bc6d-2eee6d6304df {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.431105] env[63202]: DEBUG oslo_vmware.api [None req-31c518cb-d896-429a-88e7-bb67df082013 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Waiting for the task: (returnval){ [ 695.431105] env[63202]: value = "task-1385370" [ 695.431105] env[63202]: _type = "Task" [ 695.431105] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.438987] env[63202]: DEBUG oslo_vmware.api [None req-31c518cb-d896-429a-88e7-bb67df082013 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385370, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.606034] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance fea91da2-186a-44b6-8e20-535e3b123890 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 695.941391] env[63202]: DEBUG oslo_vmware.api [None req-31c518cb-d896-429a-88e7-bb67df082013 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385370, 'name': PowerOffVM_Task, 'duration_secs': 0.12788} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.941706] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-31c518cb-d896-429a-88e7-bb67df082013 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 695.941897] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-31c518cb-d896-429a-88e7-bb67df082013 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 695.943044] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0f9c6ab9-d388-4ff1-88bd-faa770d58d78 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.973646] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-31c518cb-d896-429a-88e7-bb67df082013 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 695.973646] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-31c518cb-d896-429a-88e7-bb67df082013 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Deleting contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 695.973646] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-31c518cb-d896-429a-88e7-bb67df082013 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Deleting the datastore file [datastore1] c8d27e19-0f3c-4cc3-826b-e07b3a635e5e {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 695.973646] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eeb21b7b-0945-44ca-85f2-c39f59013747 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.981110] env[63202]: DEBUG oslo_vmware.api [None req-31c518cb-d896-429a-88e7-bb67df082013 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Waiting for the task: (returnval){ [ 695.981110] env[63202]: value = "task-1385372" [ 695.981110] env[63202]: _type = "Task" [ 695.981110] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.990580] env[63202]: DEBUG oslo_vmware.api [None req-31c518cb-d896-429a-88e7-bb67df082013 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385372, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.109947] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 8045881e-9bce-46e7-98c6-a7989f61a31e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 696.491155] env[63202]: DEBUG oslo_vmware.api [None req-31c518cb-d896-429a-88e7-bb67df082013 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385372, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.098235} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.491457] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-31c518cb-d896-429a-88e7-bb67df082013 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 696.491676] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-31c518cb-d896-429a-88e7-bb67df082013 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Deleted contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 696.491880] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-31c518cb-d896-429a-88e7-bb67df082013 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 696.492093] env[63202]: INFO nova.compute.manager [None req-31c518cb-d896-429a-88e7-bb67df082013 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Took 1.08 seconds to destroy the instance on the hypervisor. [ 696.492380] env[63202]: DEBUG oslo.service.loopingcall [None req-31c518cb-d896-429a-88e7-bb67df082013 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 696.492579] env[63202]: DEBUG nova.compute.manager [-] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 696.492700] env[63202]: DEBUG nova.network.neutron [-] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 696.507160] env[63202]: DEBUG nova.network.neutron [-] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 696.613076] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance e3a6ad78-4f46-42d4-935c-3cf310123530 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 697.009903] env[63202]: DEBUG nova.network.neutron [-] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.116083] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance c9e0ee31-de9e-420d-8dad-380391d6f4e3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 697.512277] env[63202]: INFO nova.compute.manager [-] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Took 1.02 seconds to deallocate network for instance. [ 697.618581] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance e24b7994-3786-4709-bcb5-e0b2c9731f2a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 698.018235] env[63202]: DEBUG oslo_concurrency.lockutils [None req-31c518cb-d896-429a-88e7-bb67df082013 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 698.121249] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 698.623896] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 95d2a2d7-443b-43d6-97ff-fe36fd11002f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 699.126408] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 39d9b91b-c009-40e1-ad2b-ebef650188a4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 699.629148] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 06bd8147-5479-4c5e-8ba2-6d3a205ba05e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 700.131813] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 700.635554] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 700.635798] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=63202) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 700.635940] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=63202) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 700.982752] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-207085a7-46f3-4c71-8195-171f51a87f60 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.990626] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a90dccec-6891-4601-9e2f-8e4433c470b2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.019958] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c502f63-124f-404a-bb2f-976acd231d63 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.027944] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fcf8993-3949-494c-8edc-5793d5b3ad86 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.041152] env[63202]: DEBUG nova.compute.provider_tree [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 701.546568] env[63202]: DEBUG nova.scheduler.client.report [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 702.050087] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63202) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 702.050370] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 15.546s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 702.050658] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.995s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 702.052211] env[63202]: INFO nova.compute.claims [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 703.371455] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63e2a8dc-533a-4706-b85d-813d0adc8953 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.378651] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c97e921-9cce-4c09-a71f-7e2bba441d83 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.407891] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04c9cd4e-34cb-4739-969c-8fc3006a7a81 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.414816] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4b32fd6-bded-4398-9583-5830558884df {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.427142] env[63202]: DEBUG nova.compute.provider_tree [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 703.930908] env[63202]: DEBUG nova.scheduler.client.report [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 704.437019] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.386s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 704.437554] env[63202]: DEBUG nova.compute.manager [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 704.440162] env[63202]: DEBUG oslo_concurrency.lockutils [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.116s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 704.441832] env[63202]: INFO nova.compute.claims [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 704.946606] env[63202]: DEBUG nova.compute.utils [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 704.949892] env[63202]: DEBUG nova.compute.manager [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 704.950126] env[63202]: DEBUG nova.network.neutron [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 705.005585] env[63202]: DEBUG nova.policy [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '113be6fb6fba477ba1f05e8318e270cf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '502285adbe1748179849106b302360b6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 705.356673] env[63202]: DEBUG nova.network.neutron [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] Successfully created port: 0cb05e8b-98bc-4258-bbfa-da06bc6f66cf {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 705.451350] env[63202]: DEBUG nova.compute.manager [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 705.864058] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbbc0623-8545-4f4a-9562-4d190a69f8b7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.872073] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa9e3fd1-3db9-44b2-9504-79cc8d1d9594 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.903826] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48612f4c-7306-4bfe-81a5-7a93a0648b3b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.911658] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4e4048c-b67e-4ee2-8c31-c5b660548825 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.925427] env[63202]: DEBUG nova.compute.provider_tree [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 706.009167] env[63202]: DEBUG nova.compute.manager [req-8c752507-baf9-43ea-b52b-c9c2a1483404 req-df2b063a-7cac-40e3-a326-c651f1c1da71 service nova] [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] Received event network-changed-0cb05e8b-98bc-4258-bbfa-da06bc6f66cf {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 706.009418] env[63202]: DEBUG nova.compute.manager [req-8c752507-baf9-43ea-b52b-c9c2a1483404 req-df2b063a-7cac-40e3-a326-c651f1c1da71 service nova] [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] Refreshing instance network info cache due to event network-changed-0cb05e8b-98bc-4258-bbfa-da06bc6f66cf. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 706.009743] env[63202]: DEBUG oslo_concurrency.lockutils [req-8c752507-baf9-43ea-b52b-c9c2a1483404 req-df2b063a-7cac-40e3-a326-c651f1c1da71 service nova] Acquiring lock "refresh_cache-864e1f52-f77e-4596-a5b5-dff3679f911f" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 706.009915] env[63202]: DEBUG oslo_concurrency.lockutils [req-8c752507-baf9-43ea-b52b-c9c2a1483404 req-df2b063a-7cac-40e3-a326-c651f1c1da71 service nova] Acquired lock "refresh_cache-864e1f52-f77e-4596-a5b5-dff3679f911f" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 706.010474] env[63202]: DEBUG nova.network.neutron [req-8c752507-baf9-43ea-b52b-c9c2a1483404 req-df2b063a-7cac-40e3-a326-c651f1c1da71 service nova] [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] Refreshing network info cache for port 0cb05e8b-98bc-4258-bbfa-da06bc6f66cf {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 706.205878] env[63202]: ERROR nova.compute.manager [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0cb05e8b-98bc-4258-bbfa-da06bc6f66cf, please check neutron logs for more information. [ 706.205878] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 706.205878] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 706.205878] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 706.205878] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 706.205878] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 706.205878] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 706.205878] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 706.205878] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 706.205878] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 706.205878] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 706.205878] env[63202]: ERROR nova.compute.manager raise self.value [ 706.205878] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 706.205878] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 706.205878] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 706.205878] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 706.206408] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 706.206408] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 706.206408] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0cb05e8b-98bc-4258-bbfa-da06bc6f66cf, please check neutron logs for more information. [ 706.206408] env[63202]: ERROR nova.compute.manager [ 706.206408] env[63202]: Traceback (most recent call last): [ 706.206408] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 706.206408] env[63202]: listener.cb(fileno) [ 706.206408] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 706.206408] env[63202]: result = function(*args, **kwargs) [ 706.206408] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 706.206408] env[63202]: return func(*args, **kwargs) [ 706.206408] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 706.206408] env[63202]: raise e [ 706.206408] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 706.206408] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 706.206408] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 706.206408] env[63202]: created_port_ids = self._update_ports_for_instance( [ 706.206408] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 706.206408] env[63202]: with excutils.save_and_reraise_exception(): [ 706.206408] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 706.206408] env[63202]: self.force_reraise() [ 706.206408] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 706.206408] env[63202]: raise self.value [ 706.206408] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 706.206408] env[63202]: updated_port = self._update_port( [ 706.206408] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 706.206408] env[63202]: _ensure_no_port_binding_failure(port) [ 706.206408] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 706.206408] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 706.207569] env[63202]: nova.exception.PortBindingFailed: Binding failed for port 0cb05e8b-98bc-4258-bbfa-da06bc6f66cf, please check neutron logs for more information. [ 706.207569] env[63202]: Removing descriptor: 20 [ 706.428388] env[63202]: DEBUG nova.scheduler.client.report [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 706.462876] env[63202]: DEBUG nova.compute.manager [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 706.485970] env[63202]: DEBUG nova.virt.hardware [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 706.486240] env[63202]: DEBUG nova.virt.hardware [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 706.486428] env[63202]: DEBUG nova.virt.hardware [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 706.486612] env[63202]: DEBUG nova.virt.hardware [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 706.486752] env[63202]: DEBUG nova.virt.hardware [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 706.486891] env[63202]: DEBUG nova.virt.hardware [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 706.487103] env[63202]: DEBUG nova.virt.hardware [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 706.487259] env[63202]: DEBUG nova.virt.hardware [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 706.487421] env[63202]: DEBUG nova.virt.hardware [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 706.487574] env[63202]: DEBUG nova.virt.hardware [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 706.487735] env[63202]: DEBUG nova.virt.hardware [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 706.488626] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e8c0acf-ccd3-4ecc-97a3-05eb1bcc2227 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.496811] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af333f37-a643-4766-b670-074418e89af7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.510512] env[63202]: ERROR nova.compute.manager [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0cb05e8b-98bc-4258-bbfa-da06bc6f66cf, please check neutron logs for more information. [ 706.510512] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] Traceback (most recent call last): [ 706.510512] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 706.510512] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] yield resources [ 706.510512] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 706.510512] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] self.driver.spawn(context, instance, image_meta, [ 706.510512] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 706.510512] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 706.510512] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 706.510512] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] vm_ref = self.build_virtual_machine(instance, [ 706.510512] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 706.510927] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] vif_infos = vmwarevif.get_vif_info(self._session, [ 706.510927] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 706.510927] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] for vif in network_info: [ 706.510927] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 706.510927] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] return self._sync_wrapper(fn, *args, **kwargs) [ 706.510927] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 706.510927] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] self.wait() [ 706.510927] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 706.510927] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] self[:] = self._gt.wait() [ 706.510927] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 706.510927] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] return self._exit_event.wait() [ 706.510927] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 706.510927] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] current.throw(*self._exc) [ 706.511388] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 706.511388] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] result = function(*args, **kwargs) [ 706.511388] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 706.511388] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] return func(*args, **kwargs) [ 706.511388] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 706.511388] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] raise e [ 706.511388] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 706.511388] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] nwinfo = self.network_api.allocate_for_instance( [ 706.511388] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 706.511388] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] created_port_ids = self._update_ports_for_instance( [ 706.511388] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 706.511388] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] with excutils.save_and_reraise_exception(): [ 706.511388] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 706.511805] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] self.force_reraise() [ 706.511805] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 706.511805] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] raise self.value [ 706.511805] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 706.511805] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] updated_port = self._update_port( [ 706.511805] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 706.511805] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] _ensure_no_port_binding_failure(port) [ 706.511805] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 706.511805] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] raise exception.PortBindingFailed(port_id=port['id']) [ 706.511805] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] nova.exception.PortBindingFailed: Binding failed for port 0cb05e8b-98bc-4258-bbfa-da06bc6f66cf, please check neutron logs for more information. [ 706.511805] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] [ 706.511805] env[63202]: INFO nova.compute.manager [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] Terminating instance [ 706.514034] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] Acquiring lock "refresh_cache-864e1f52-f77e-4596-a5b5-dff3679f911f" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 706.526825] env[63202]: DEBUG nova.network.neutron [req-8c752507-baf9-43ea-b52b-c9c2a1483404 req-df2b063a-7cac-40e3-a326-c651f1c1da71 service nova] [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 706.578853] env[63202]: DEBUG nova.network.neutron [req-8c752507-baf9-43ea-b52b-c9c2a1483404 req-df2b063a-7cac-40e3-a326-c651f1c1da71 service nova] [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.935977] env[63202]: DEBUG oslo_concurrency.lockutils [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.493s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.935977] env[63202]: DEBUG nova.compute.manager [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 706.937318] env[63202]: DEBUG oslo_concurrency.lockutils [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.213s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.938966] env[63202]: INFO nova.compute.claims [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] [instance: de5f9555-3959-41cd-b63a-db128a001631] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 707.082255] env[63202]: DEBUG oslo_concurrency.lockutils [req-8c752507-baf9-43ea-b52b-c9c2a1483404 req-df2b063a-7cac-40e3-a326-c651f1c1da71 service nova] Releasing lock "refresh_cache-864e1f52-f77e-4596-a5b5-dff3679f911f" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 707.082656] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] Acquired lock "refresh_cache-864e1f52-f77e-4596-a5b5-dff3679f911f" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.082837] env[63202]: DEBUG nova.network.neutron [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 707.448143] env[63202]: DEBUG nova.compute.utils [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 707.449114] env[63202]: DEBUG nova.compute.manager [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 707.449114] env[63202]: DEBUG nova.network.neutron [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 707.503715] env[63202]: DEBUG nova.policy [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '80d5f12ec7d24d3eb026057278a2ab9f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ec27a05773614d3a9df8c3921ec2f43d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 707.601717] env[63202]: DEBUG nova.network.neutron [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 707.654440] env[63202]: DEBUG nova.network.neutron [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.805692] env[63202]: DEBUG nova.network.neutron [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] Successfully created port: 3cb97f6c-facd-4990-8a29-63073f5d22a7 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 707.956291] env[63202]: DEBUG nova.compute.manager [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 708.037810] env[63202]: DEBUG nova.compute.manager [req-d49f1a19-5be3-4616-adbc-c6fd748cdd6f req-8409097c-7256-43d8-bd26-44d2dca8b9ee service nova] [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] Received event network-vif-deleted-0cb05e8b-98bc-4258-bbfa-da06bc6f66cf {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 708.159540] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] Releasing lock "refresh_cache-864e1f52-f77e-4596-a5b5-dff3679f911f" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.159931] env[63202]: DEBUG nova.compute.manager [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 708.160155] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 708.160644] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9cf07d0b-1506-4bae-b461-cd34aa744b0f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.169664] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ac511af-fec6-4ee1-9718-0190c66cd2ac {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.203742] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 864e1f52-f77e-4596-a5b5-dff3679f911f could not be found. [ 708.203742] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 708.203742] env[63202]: INFO nova.compute.manager [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 708.203742] env[63202]: DEBUG oslo.service.loopingcall [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 708.205706] env[63202]: DEBUG nova.compute.manager [-] [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 708.205778] env[63202]: DEBUG nova.network.neutron [-] [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 708.228538] env[63202]: DEBUG nova.network.neutron [-] [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 708.502718] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32d5dc90-1938-443c-a720-7734908699e4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.510516] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78eea5ea-d7dd-4485-a3ff-4ff1f26fffde {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.542244] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adb3ec50-af60-4cb4-a2aa-b5058201deed {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.550530] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2d4c7fd-7693-4a8c-9a64-fbb503551773 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.565013] env[63202]: DEBUG nova.compute.provider_tree [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 708.629481] env[63202]: ERROR nova.compute.manager [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3cb97f6c-facd-4990-8a29-63073f5d22a7, please check neutron logs for more information. [ 708.629481] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 708.629481] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 708.629481] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 708.629481] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 708.629481] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 708.629481] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 708.629481] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 708.629481] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 708.629481] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 708.629481] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 708.629481] env[63202]: ERROR nova.compute.manager raise self.value [ 708.629481] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 708.629481] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 708.629481] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 708.629481] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 708.629955] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 708.629955] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 708.629955] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3cb97f6c-facd-4990-8a29-63073f5d22a7, please check neutron logs for more information. [ 708.629955] env[63202]: ERROR nova.compute.manager [ 708.629955] env[63202]: Traceback (most recent call last): [ 708.629955] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 708.629955] env[63202]: listener.cb(fileno) [ 708.629955] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 708.629955] env[63202]: result = function(*args, **kwargs) [ 708.629955] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 708.629955] env[63202]: return func(*args, **kwargs) [ 708.629955] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 708.629955] env[63202]: raise e [ 708.629955] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 708.629955] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 708.629955] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 708.629955] env[63202]: created_port_ids = self._update_ports_for_instance( [ 708.629955] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 708.629955] env[63202]: with excutils.save_and_reraise_exception(): [ 708.629955] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 708.629955] env[63202]: self.force_reraise() [ 708.629955] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 708.629955] env[63202]: raise self.value [ 708.629955] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 708.629955] env[63202]: updated_port = self._update_port( [ 708.629955] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 708.629955] env[63202]: _ensure_no_port_binding_failure(port) [ 708.629955] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 708.629955] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 708.630842] env[63202]: nova.exception.PortBindingFailed: Binding failed for port 3cb97f6c-facd-4990-8a29-63073f5d22a7, please check neutron logs for more information. [ 708.630842] env[63202]: Removing descriptor: 20 [ 708.733232] env[63202]: DEBUG nova.network.neutron [-] [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.967259] env[63202]: DEBUG nova.compute.manager [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 708.994838] env[63202]: DEBUG nova.virt.hardware [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 708.995253] env[63202]: DEBUG nova.virt.hardware [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 708.995453] env[63202]: DEBUG nova.virt.hardware [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 708.995707] env[63202]: DEBUG nova.virt.hardware [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 708.995981] env[63202]: DEBUG nova.virt.hardware [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 708.996189] env[63202]: DEBUG nova.virt.hardware [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 708.996693] env[63202]: DEBUG nova.virt.hardware [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 708.996732] env[63202]: DEBUG nova.virt.hardware [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 708.996917] env[63202]: DEBUG nova.virt.hardware [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 708.997121] env[63202]: DEBUG nova.virt.hardware [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 708.997335] env[63202]: DEBUG nova.virt.hardware [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 708.998297] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d35c0b1e-fd54-4b88-b030-6471d6548b2d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.007477] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11675cb4-8275-4e39-89c8-95e8836adf72 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.022497] env[63202]: ERROR nova.compute.manager [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3cb97f6c-facd-4990-8a29-63073f5d22a7, please check neutron logs for more information. [ 709.022497] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] Traceback (most recent call last): [ 709.022497] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 709.022497] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] yield resources [ 709.022497] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 709.022497] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] self.driver.spawn(context, instance, image_meta, [ 709.022497] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 709.022497] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] self._vmops.spawn(context, instance, image_meta, injected_files, [ 709.022497] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 709.022497] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] vm_ref = self.build_virtual_machine(instance, [ 709.022497] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 709.022943] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] vif_infos = vmwarevif.get_vif_info(self._session, [ 709.022943] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 709.022943] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] for vif in network_info: [ 709.022943] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 709.022943] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] return self._sync_wrapper(fn, *args, **kwargs) [ 709.022943] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 709.022943] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] self.wait() [ 709.022943] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 709.022943] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] self[:] = self._gt.wait() [ 709.022943] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 709.022943] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] return self._exit_event.wait() [ 709.022943] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 709.022943] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] current.throw(*self._exc) [ 709.023318] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 709.023318] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] result = function(*args, **kwargs) [ 709.023318] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 709.023318] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] return func(*args, **kwargs) [ 709.023318] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 709.023318] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] raise e [ 709.023318] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 709.023318] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] nwinfo = self.network_api.allocate_for_instance( [ 709.023318] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 709.023318] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] created_port_ids = self._update_ports_for_instance( [ 709.023318] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 709.023318] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] with excutils.save_and_reraise_exception(): [ 709.023318] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 709.023694] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] self.force_reraise() [ 709.023694] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 709.023694] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] raise self.value [ 709.023694] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 709.023694] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] updated_port = self._update_port( [ 709.023694] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 709.023694] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] _ensure_no_port_binding_failure(port) [ 709.023694] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 709.023694] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] raise exception.PortBindingFailed(port_id=port['id']) [ 709.023694] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] nova.exception.PortBindingFailed: Binding failed for port 3cb97f6c-facd-4990-8a29-63073f5d22a7, please check neutron logs for more information. [ 709.023694] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] [ 709.023694] env[63202]: INFO nova.compute.manager [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] Terminating instance [ 709.024797] env[63202]: DEBUG oslo_concurrency.lockutils [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] Acquiring lock "refresh_cache-eff8282a-b3cd-4038-9f48-94cd4c4afc55" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 709.024948] env[63202]: DEBUG oslo_concurrency.lockutils [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] Acquired lock "refresh_cache-eff8282a-b3cd-4038-9f48-94cd4c4afc55" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.025128] env[63202]: DEBUG nova.network.neutron [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 709.066956] env[63202]: DEBUG nova.scheduler.client.report [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 709.235702] env[63202]: INFO nova.compute.manager [-] [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] Took 1.03 seconds to deallocate network for instance. [ 709.238132] env[63202]: DEBUG nova.compute.claims [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 709.238375] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 709.544021] env[63202]: DEBUG nova.network.neutron [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 709.572338] env[63202]: DEBUG oslo_concurrency.lockutils [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.635s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 709.572829] env[63202]: DEBUG nova.compute.manager [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] [instance: de5f9555-3959-41cd-b63a-db128a001631] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 709.575447] env[63202]: DEBUG oslo_concurrency.lockutils [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.986s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 709.576965] env[63202]: INFO nova.compute.claims [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] [instance: 2f87939c-d615-4d13-8695-971e83ca7843] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 709.597479] env[63202]: DEBUG nova.network.neutron [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.079984] env[63202]: DEBUG nova.compute.manager [req-4b90709d-8dc2-490f-8aa2-1ff941f90172 req-87a3110e-8de5-4920-a8df-6a6c0195259d service nova] [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] Received event network-changed-3cb97f6c-facd-4990-8a29-63073f5d22a7 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 710.080320] env[63202]: DEBUG nova.compute.manager [req-4b90709d-8dc2-490f-8aa2-1ff941f90172 req-87a3110e-8de5-4920-a8df-6a6c0195259d service nova] [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] Refreshing instance network info cache due to event network-changed-3cb97f6c-facd-4990-8a29-63073f5d22a7. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 710.080499] env[63202]: DEBUG oslo_concurrency.lockutils [req-4b90709d-8dc2-490f-8aa2-1ff941f90172 req-87a3110e-8de5-4920-a8df-6a6c0195259d service nova] Acquiring lock "refresh_cache-eff8282a-b3cd-4038-9f48-94cd4c4afc55" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.082056] env[63202]: DEBUG nova.compute.utils [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 710.085681] env[63202]: DEBUG nova.compute.manager [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] [instance: de5f9555-3959-41cd-b63a-db128a001631] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 710.085848] env[63202]: DEBUG nova.network.neutron [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] [instance: de5f9555-3959-41cd-b63a-db128a001631] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 710.100337] env[63202]: DEBUG oslo_concurrency.lockutils [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] Releasing lock "refresh_cache-eff8282a-b3cd-4038-9f48-94cd4c4afc55" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 710.100460] env[63202]: DEBUG nova.compute.manager [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 710.100628] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 710.100922] env[63202]: DEBUG oslo_concurrency.lockutils [req-4b90709d-8dc2-490f-8aa2-1ff941f90172 req-87a3110e-8de5-4920-a8df-6a6c0195259d service nova] Acquired lock "refresh_cache-eff8282a-b3cd-4038-9f48-94cd4c4afc55" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.101100] env[63202]: DEBUG nova.network.neutron [req-4b90709d-8dc2-490f-8aa2-1ff941f90172 req-87a3110e-8de5-4920-a8df-6a6c0195259d service nova] [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] Refreshing network info cache for port 3cb97f6c-facd-4990-8a29-63073f5d22a7 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 710.102021] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-936622eb-7f77-4c2c-b7a2-f14229db84b2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.112802] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fb9cb32-b77a-430b-84e6-b66ddf1e78ee {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.125144] env[63202]: DEBUG nova.policy [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '27911ffe75c94f27a57b83460958aac8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '838693f8fbad46f4b52cb0813f416f99', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 710.134796] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance eff8282a-b3cd-4038-9f48-94cd4c4afc55 could not be found. [ 710.135016] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 710.135210] env[63202]: INFO nova.compute.manager [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] Took 0.03 seconds to destroy the instance on the hypervisor. [ 710.135448] env[63202]: DEBUG oslo.service.loopingcall [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 710.135655] env[63202]: DEBUG nova.compute.manager [-] [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 710.135742] env[63202]: DEBUG nova.network.neutron [-] [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 710.149289] env[63202]: DEBUG nova.network.neutron [-] [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 710.378494] env[63202]: DEBUG nova.network.neutron [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] [instance: de5f9555-3959-41cd-b63a-db128a001631] Successfully created port: e280fa17-a568-4c0b-9e68-74717594a05a {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 710.586714] env[63202]: DEBUG nova.compute.manager [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] [instance: de5f9555-3959-41cd-b63a-db128a001631] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 710.629034] env[63202]: DEBUG nova.network.neutron [req-4b90709d-8dc2-490f-8aa2-1ff941f90172 req-87a3110e-8de5-4920-a8df-6a6c0195259d service nova] [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 710.651074] env[63202]: DEBUG nova.network.neutron [-] [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.708870] env[63202]: DEBUG nova.network.neutron [req-4b90709d-8dc2-490f-8aa2-1ff941f90172 req-87a3110e-8de5-4920-a8df-6a6c0195259d service nova] [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.967684] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9280d620-9864-4a12-9283-7223470317b8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.975315] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dde63c2a-366a-44ce-aeec-b0649707c585 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.006009] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b77ada3-73e4-4fee-8099-b29eaed2671a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.013381] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1af7a887-f7e9-4d01-8848-18ff3cfec765 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.026749] env[63202]: DEBUG nova.compute.provider_tree [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 711.153414] env[63202]: INFO nova.compute.manager [-] [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] Took 1.02 seconds to deallocate network for instance. [ 711.155460] env[63202]: DEBUG nova.compute.claims [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 711.155632] env[63202]: DEBUG oslo_concurrency.lockutils [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 711.213339] env[63202]: DEBUG oslo_concurrency.lockutils [req-4b90709d-8dc2-490f-8aa2-1ff941f90172 req-87a3110e-8de5-4920-a8df-6a6c0195259d service nova] Releasing lock "refresh_cache-eff8282a-b3cd-4038-9f48-94cd4c4afc55" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 711.213696] env[63202]: DEBUG nova.compute.manager [req-4b90709d-8dc2-490f-8aa2-1ff941f90172 req-87a3110e-8de5-4920-a8df-6a6c0195259d service nova] [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] Received event network-vif-deleted-3cb97f6c-facd-4990-8a29-63073f5d22a7 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 711.350085] env[63202]: ERROR nova.compute.manager [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e280fa17-a568-4c0b-9e68-74717594a05a, please check neutron logs for more information. [ 711.350085] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 711.350085] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 711.350085] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 711.350085] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 711.350085] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 711.350085] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 711.350085] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 711.350085] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 711.350085] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 711.350085] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 711.350085] env[63202]: ERROR nova.compute.manager raise self.value [ 711.350085] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 711.350085] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 711.350085] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 711.350085] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 711.350631] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 711.350631] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 711.350631] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e280fa17-a568-4c0b-9e68-74717594a05a, please check neutron logs for more information. [ 711.350631] env[63202]: ERROR nova.compute.manager [ 711.350631] env[63202]: Traceback (most recent call last): [ 711.350631] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 711.350631] env[63202]: listener.cb(fileno) [ 711.350631] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 711.350631] env[63202]: result = function(*args, **kwargs) [ 711.350631] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 711.350631] env[63202]: return func(*args, **kwargs) [ 711.350631] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 711.350631] env[63202]: raise e [ 711.350631] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 711.350631] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 711.350631] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 711.350631] env[63202]: created_port_ids = self._update_ports_for_instance( [ 711.350631] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 711.350631] env[63202]: with excutils.save_and_reraise_exception(): [ 711.350631] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 711.350631] env[63202]: self.force_reraise() [ 711.350631] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 711.350631] env[63202]: raise self.value [ 711.350631] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 711.350631] env[63202]: updated_port = self._update_port( [ 711.350631] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 711.350631] env[63202]: _ensure_no_port_binding_failure(port) [ 711.350631] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 711.350631] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 711.351946] env[63202]: nova.exception.PortBindingFailed: Binding failed for port e280fa17-a568-4c0b-9e68-74717594a05a, please check neutron logs for more information. [ 711.351946] env[63202]: Removing descriptor: 20 [ 711.532725] env[63202]: DEBUG nova.scheduler.client.report [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 711.599348] env[63202]: DEBUG nova.compute.manager [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] [instance: de5f9555-3959-41cd-b63a-db128a001631] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 711.623724] env[63202]: DEBUG nova.virt.hardware [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 711.623969] env[63202]: DEBUG nova.virt.hardware [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 711.624139] env[63202]: DEBUG nova.virt.hardware [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 711.624321] env[63202]: DEBUG nova.virt.hardware [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 711.624461] env[63202]: DEBUG nova.virt.hardware [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 711.624602] env[63202]: DEBUG nova.virt.hardware [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 711.624799] env[63202]: DEBUG nova.virt.hardware [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 711.624950] env[63202]: DEBUG nova.virt.hardware [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 711.625126] env[63202]: DEBUG nova.virt.hardware [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 711.625284] env[63202]: DEBUG nova.virt.hardware [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 711.625497] env[63202]: DEBUG nova.virt.hardware [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 711.626349] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-115836f8-5515-4d91-918c-6b4f66f72ea7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.634178] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bdc13b9-2e43-4913-99a0-49740ee239ae {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.647833] env[63202]: ERROR nova.compute.manager [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] [instance: de5f9555-3959-41cd-b63a-db128a001631] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e280fa17-a568-4c0b-9e68-74717594a05a, please check neutron logs for more information. [ 711.647833] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] Traceback (most recent call last): [ 711.647833] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 711.647833] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] yield resources [ 711.647833] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 711.647833] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] self.driver.spawn(context, instance, image_meta, [ 711.647833] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 711.647833] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] self._vmops.spawn(context, instance, image_meta, injected_files, [ 711.647833] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 711.647833] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] vm_ref = self.build_virtual_machine(instance, [ 711.647833] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 711.648329] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] vif_infos = vmwarevif.get_vif_info(self._session, [ 711.648329] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 711.648329] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] for vif in network_info: [ 711.648329] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 711.648329] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] return self._sync_wrapper(fn, *args, **kwargs) [ 711.648329] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 711.648329] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] self.wait() [ 711.648329] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 711.648329] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] self[:] = self._gt.wait() [ 711.648329] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 711.648329] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] return self._exit_event.wait() [ 711.648329] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 711.648329] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] current.throw(*self._exc) [ 711.648697] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 711.648697] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] result = function(*args, **kwargs) [ 711.648697] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 711.648697] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] return func(*args, **kwargs) [ 711.648697] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 711.648697] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] raise e [ 711.648697] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 711.648697] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] nwinfo = self.network_api.allocate_for_instance( [ 711.648697] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 711.648697] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] created_port_ids = self._update_ports_for_instance( [ 711.648697] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 711.648697] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] with excutils.save_and_reraise_exception(): [ 711.648697] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 711.649022] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] self.force_reraise() [ 711.649022] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 711.649022] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] raise self.value [ 711.649022] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 711.649022] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] updated_port = self._update_port( [ 711.649022] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 711.649022] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] _ensure_no_port_binding_failure(port) [ 711.649022] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 711.649022] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] raise exception.PortBindingFailed(port_id=port['id']) [ 711.649022] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] nova.exception.PortBindingFailed: Binding failed for port e280fa17-a568-4c0b-9e68-74717594a05a, please check neutron logs for more information. [ 711.649022] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] [ 711.649022] env[63202]: INFO nova.compute.manager [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] [instance: de5f9555-3959-41cd-b63a-db128a001631] Terminating instance [ 711.650272] env[63202]: DEBUG oslo_concurrency.lockutils [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] Acquiring lock "refresh_cache-de5f9555-3959-41cd-b63a-db128a001631" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 711.650430] env[63202]: DEBUG oslo_concurrency.lockutils [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] Acquired lock "refresh_cache-de5f9555-3959-41cd-b63a-db128a001631" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.650590] env[63202]: DEBUG nova.network.neutron [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] [instance: de5f9555-3959-41cd-b63a-db128a001631] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 712.037474] env[63202]: DEBUG oslo_concurrency.lockutils [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.462s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.038094] env[63202]: DEBUG nova.compute.manager [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] [instance: 2f87939c-d615-4d13-8695-971e83ca7843] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 712.041124] env[63202]: DEBUG oslo_concurrency.lockutils [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.441s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 712.042341] env[63202]: INFO nova.compute.claims [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 712.105836] env[63202]: DEBUG nova.compute.manager [req-9f270464-90f6-4b8c-97f3-525aa7c889bb req-8117deff-75b0-4dff-a3e5-aded5c84a00e service nova] [instance: de5f9555-3959-41cd-b63a-db128a001631] Received event network-changed-e280fa17-a568-4c0b-9e68-74717594a05a {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 712.106062] env[63202]: DEBUG nova.compute.manager [req-9f270464-90f6-4b8c-97f3-525aa7c889bb req-8117deff-75b0-4dff-a3e5-aded5c84a00e service nova] [instance: de5f9555-3959-41cd-b63a-db128a001631] Refreshing instance network info cache due to event network-changed-e280fa17-a568-4c0b-9e68-74717594a05a. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 712.106289] env[63202]: DEBUG oslo_concurrency.lockutils [req-9f270464-90f6-4b8c-97f3-525aa7c889bb req-8117deff-75b0-4dff-a3e5-aded5c84a00e service nova] Acquiring lock "refresh_cache-de5f9555-3959-41cd-b63a-db128a001631" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.168969] env[63202]: DEBUG nova.network.neutron [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] [instance: de5f9555-3959-41cd-b63a-db128a001631] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 712.248893] env[63202]: DEBUG nova.network.neutron [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] [instance: de5f9555-3959-41cd-b63a-db128a001631] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.549041] env[63202]: DEBUG nova.compute.utils [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 712.550026] env[63202]: DEBUG nova.compute.manager [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] [instance: 2f87939c-d615-4d13-8695-971e83ca7843] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 712.550122] env[63202]: DEBUG nova.network.neutron [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] [instance: 2f87939c-d615-4d13-8695-971e83ca7843] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 712.589389] env[63202]: DEBUG nova.policy [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ed7c675e4ea04069be4075f38014c3f6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1e6b17873d3742288def03405a4180ac', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 712.751663] env[63202]: DEBUG oslo_concurrency.lockutils [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] Releasing lock "refresh_cache-de5f9555-3959-41cd-b63a-db128a001631" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 712.752104] env[63202]: DEBUG nova.compute.manager [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] [instance: de5f9555-3959-41cd-b63a-db128a001631] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 712.752295] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] [instance: de5f9555-3959-41cd-b63a-db128a001631] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 712.752605] env[63202]: DEBUG oslo_concurrency.lockutils [req-9f270464-90f6-4b8c-97f3-525aa7c889bb req-8117deff-75b0-4dff-a3e5-aded5c84a00e service nova] Acquired lock "refresh_cache-de5f9555-3959-41cd-b63a-db128a001631" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.752771] env[63202]: DEBUG nova.network.neutron [req-9f270464-90f6-4b8c-97f3-525aa7c889bb req-8117deff-75b0-4dff-a3e5-aded5c84a00e service nova] [instance: de5f9555-3959-41cd-b63a-db128a001631] Refreshing network info cache for port e280fa17-a568-4c0b-9e68-74717594a05a {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 712.754018] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2ad0fe15-6dea-421e-9ebc-f7b08be39fbf {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.765716] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7866b3d-302e-4d6a-966b-7025d29e7cdc {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.787928] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] [instance: de5f9555-3959-41cd-b63a-db128a001631] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance de5f9555-3959-41cd-b63a-db128a001631 could not be found. [ 712.788193] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] [instance: de5f9555-3959-41cd-b63a-db128a001631] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 712.788389] env[63202]: INFO nova.compute.manager [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] [instance: de5f9555-3959-41cd-b63a-db128a001631] Took 0.04 seconds to destroy the instance on the hypervisor. [ 712.788631] env[63202]: DEBUG oslo.service.loopingcall [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 712.788855] env[63202]: DEBUG nova.compute.manager [-] [instance: de5f9555-3959-41cd-b63a-db128a001631] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 712.788945] env[63202]: DEBUG nova.network.neutron [-] [instance: de5f9555-3959-41cd-b63a-db128a001631] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 712.803647] env[63202]: DEBUG nova.network.neutron [-] [instance: de5f9555-3959-41cd-b63a-db128a001631] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 712.846989] env[63202]: DEBUG nova.network.neutron [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] [instance: 2f87939c-d615-4d13-8695-971e83ca7843] Successfully created port: 232b0dca-d02a-4796-921b-b5dca87b0a1b {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 713.053456] env[63202]: DEBUG nova.compute.manager [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] [instance: 2f87939c-d615-4d13-8695-971e83ca7843] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 713.306478] env[63202]: DEBUG nova.network.neutron [-] [instance: de5f9555-3959-41cd-b63a-db128a001631] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.391983] env[63202]: DEBUG nova.network.neutron [req-9f270464-90f6-4b8c-97f3-525aa7c889bb req-8117deff-75b0-4dff-a3e5-aded5c84a00e service nova] [instance: de5f9555-3959-41cd-b63a-db128a001631] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 713.437527] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87f2eb8e-3694-4628-8216-34303edc15bf {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.445395] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aee4c1db-e145-4a5d-9645-47f2512a9e03 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.478592] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f88ce0d-1908-434f-b854-3adad4ea1b64 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.486113] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-669d4bc7-ee17-4007-83ab-9298461033cc {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.502905] env[63202]: DEBUG nova.compute.provider_tree [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 713.538876] env[63202]: DEBUG nova.network.neutron [req-9f270464-90f6-4b8c-97f3-525aa7c889bb req-8117deff-75b0-4dff-a3e5-aded5c84a00e service nova] [instance: de5f9555-3959-41cd-b63a-db128a001631] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.686393] env[63202]: ERROR nova.compute.manager [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 232b0dca-d02a-4796-921b-b5dca87b0a1b, please check neutron logs for more information. [ 713.686393] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 713.686393] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 713.686393] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 713.686393] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 713.686393] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 713.686393] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 713.686393] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 713.686393] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 713.686393] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 713.686393] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 713.686393] env[63202]: ERROR nova.compute.manager raise self.value [ 713.686393] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 713.686393] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 713.686393] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 713.686393] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 713.687015] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 713.687015] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 713.687015] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 232b0dca-d02a-4796-921b-b5dca87b0a1b, please check neutron logs for more information. [ 713.687015] env[63202]: ERROR nova.compute.manager [ 713.687015] env[63202]: Traceback (most recent call last): [ 713.687015] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 713.687015] env[63202]: listener.cb(fileno) [ 713.687015] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 713.687015] env[63202]: result = function(*args, **kwargs) [ 713.687015] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 713.687015] env[63202]: return func(*args, **kwargs) [ 713.687015] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 713.687015] env[63202]: raise e [ 713.687015] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 713.687015] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 713.687015] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 713.687015] env[63202]: created_port_ids = self._update_ports_for_instance( [ 713.687015] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 713.687015] env[63202]: with excutils.save_and_reraise_exception(): [ 713.687015] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 713.687015] env[63202]: self.force_reraise() [ 713.687015] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 713.687015] env[63202]: raise self.value [ 713.687015] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 713.687015] env[63202]: updated_port = self._update_port( [ 713.687015] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 713.687015] env[63202]: _ensure_no_port_binding_failure(port) [ 713.687015] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 713.687015] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 713.688291] env[63202]: nova.exception.PortBindingFailed: Binding failed for port 232b0dca-d02a-4796-921b-b5dca87b0a1b, please check neutron logs for more information. [ 713.688291] env[63202]: Removing descriptor: 20 [ 713.808913] env[63202]: INFO nova.compute.manager [-] [instance: de5f9555-3959-41cd-b63a-db128a001631] Took 1.02 seconds to deallocate network for instance. [ 713.811260] env[63202]: DEBUG nova.compute.claims [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] [instance: de5f9555-3959-41cd-b63a-db128a001631] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 713.811431] env[63202]: DEBUG oslo_concurrency.lockutils [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 714.006640] env[63202]: DEBUG nova.scheduler.client.report [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 714.041861] env[63202]: DEBUG oslo_concurrency.lockutils [req-9f270464-90f6-4b8c-97f3-525aa7c889bb req-8117deff-75b0-4dff-a3e5-aded5c84a00e service nova] Releasing lock "refresh_cache-de5f9555-3959-41cd-b63a-db128a001631" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.042102] env[63202]: DEBUG nova.compute.manager [req-9f270464-90f6-4b8c-97f3-525aa7c889bb req-8117deff-75b0-4dff-a3e5-aded5c84a00e service nova] [instance: de5f9555-3959-41cd-b63a-db128a001631] Received event network-vif-deleted-e280fa17-a568-4c0b-9e68-74717594a05a {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 714.061904] env[63202]: DEBUG nova.compute.manager [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] [instance: 2f87939c-d615-4d13-8695-971e83ca7843] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 714.086612] env[63202]: DEBUG nova.virt.hardware [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 714.086848] env[63202]: DEBUG nova.virt.hardware [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 714.086999] env[63202]: DEBUG nova.virt.hardware [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 714.087191] env[63202]: DEBUG nova.virt.hardware [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 714.087332] env[63202]: DEBUG nova.virt.hardware [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 714.087478] env[63202]: DEBUG nova.virt.hardware [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 714.087676] env[63202]: DEBUG nova.virt.hardware [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 714.087828] env[63202]: DEBUG nova.virt.hardware [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 714.087987] env[63202]: DEBUG nova.virt.hardware [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 714.088161] env[63202]: DEBUG nova.virt.hardware [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 714.088383] env[63202]: DEBUG nova.virt.hardware [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 714.089530] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caafa9c2-0d7b-45f0-b526-e72deb0e5c91 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.097296] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-081ebdc1-e11a-4896-80e5-abc09172ec36 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.110559] env[63202]: ERROR nova.compute.manager [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] [instance: 2f87939c-d615-4d13-8695-971e83ca7843] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 232b0dca-d02a-4796-921b-b5dca87b0a1b, please check neutron logs for more information. [ 714.110559] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] Traceback (most recent call last): [ 714.110559] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 714.110559] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] yield resources [ 714.110559] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 714.110559] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] self.driver.spawn(context, instance, image_meta, [ 714.110559] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 714.110559] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] self._vmops.spawn(context, instance, image_meta, injected_files, [ 714.110559] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 714.110559] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] vm_ref = self.build_virtual_machine(instance, [ 714.110559] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 714.111111] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] vif_infos = vmwarevif.get_vif_info(self._session, [ 714.111111] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 714.111111] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] for vif in network_info: [ 714.111111] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 714.111111] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] return self._sync_wrapper(fn, *args, **kwargs) [ 714.111111] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 714.111111] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] self.wait() [ 714.111111] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 714.111111] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] self[:] = self._gt.wait() [ 714.111111] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 714.111111] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] return self._exit_event.wait() [ 714.111111] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 714.111111] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] current.throw(*self._exc) [ 714.111640] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 714.111640] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] result = function(*args, **kwargs) [ 714.111640] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 714.111640] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] return func(*args, **kwargs) [ 714.111640] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 714.111640] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] raise e [ 714.111640] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 714.111640] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] nwinfo = self.network_api.allocate_for_instance( [ 714.111640] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 714.111640] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] created_port_ids = self._update_ports_for_instance( [ 714.111640] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 714.111640] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] with excutils.save_and_reraise_exception(): [ 714.111640] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 714.111950] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] self.force_reraise() [ 714.111950] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 714.111950] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] raise self.value [ 714.111950] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 714.111950] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] updated_port = self._update_port( [ 714.111950] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 714.111950] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] _ensure_no_port_binding_failure(port) [ 714.111950] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 714.111950] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] raise exception.PortBindingFailed(port_id=port['id']) [ 714.111950] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] nova.exception.PortBindingFailed: Binding failed for port 232b0dca-d02a-4796-921b-b5dca87b0a1b, please check neutron logs for more information. [ 714.111950] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] [ 714.111950] env[63202]: INFO nova.compute.manager [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] [instance: 2f87939c-d615-4d13-8695-971e83ca7843] Terminating instance [ 714.112874] env[63202]: DEBUG oslo_concurrency.lockutils [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] Acquiring lock "refresh_cache-2f87939c-d615-4d13-8695-971e83ca7843" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 714.113173] env[63202]: DEBUG oslo_concurrency.lockutils [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] Acquired lock "refresh_cache-2f87939c-d615-4d13-8695-971e83ca7843" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.113173] env[63202]: DEBUG nova.network.neutron [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] [instance: 2f87939c-d615-4d13-8695-971e83ca7843] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 714.130341] env[63202]: DEBUG nova.compute.manager [req-5c968337-ef92-4d43-94e4-6706189167c0 req-04c668f6-50e1-4d30-b924-f5f837a26d78 service nova] [instance: 2f87939c-d615-4d13-8695-971e83ca7843] Received event network-changed-232b0dca-d02a-4796-921b-b5dca87b0a1b {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 714.130712] env[63202]: DEBUG nova.compute.manager [req-5c968337-ef92-4d43-94e4-6706189167c0 req-04c668f6-50e1-4d30-b924-f5f837a26d78 service nova] [instance: 2f87939c-d615-4d13-8695-971e83ca7843] Refreshing instance network info cache due to event network-changed-232b0dca-d02a-4796-921b-b5dca87b0a1b. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 714.130894] env[63202]: DEBUG oslo_concurrency.lockutils [req-5c968337-ef92-4d43-94e4-6706189167c0 req-04c668f6-50e1-4d30-b924-f5f837a26d78 service nova] Acquiring lock "refresh_cache-2f87939c-d615-4d13-8695-971e83ca7843" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 714.511749] env[63202]: DEBUG oslo_concurrency.lockutils [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.471s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.512280] env[63202]: DEBUG nova.compute.manager [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 714.515103] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.338s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 714.516492] env[63202]: INFO nova.compute.claims [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 06496aad-b025-48c2-8436-03d43d1ac899] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 714.637327] env[63202]: DEBUG nova.network.neutron [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] [instance: 2f87939c-d615-4d13-8695-971e83ca7843] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 714.717612] env[63202]: DEBUG nova.network.neutron [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] [instance: 2f87939c-d615-4d13-8695-971e83ca7843] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.023499] env[63202]: DEBUG nova.compute.utils [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 715.024360] env[63202]: DEBUG nova.compute.manager [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 715.024532] env[63202]: DEBUG nova.network.neutron [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 715.066937] env[63202]: DEBUG nova.policy [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '24878273948f4c2584621b6eb3f79f5a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a8318d94cc5048de936eb06da4642545', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 715.220445] env[63202]: DEBUG oslo_concurrency.lockutils [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] Releasing lock "refresh_cache-2f87939c-d615-4d13-8695-971e83ca7843" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 715.220903] env[63202]: DEBUG nova.compute.manager [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] [instance: 2f87939c-d615-4d13-8695-971e83ca7843] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 715.221074] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] [instance: 2f87939c-d615-4d13-8695-971e83ca7843] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 715.221403] env[63202]: DEBUG oslo_concurrency.lockutils [req-5c968337-ef92-4d43-94e4-6706189167c0 req-04c668f6-50e1-4d30-b924-f5f837a26d78 service nova] Acquired lock "refresh_cache-2f87939c-d615-4d13-8695-971e83ca7843" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 715.221617] env[63202]: DEBUG nova.network.neutron [req-5c968337-ef92-4d43-94e4-6706189167c0 req-04c668f6-50e1-4d30-b924-f5f837a26d78 service nova] [instance: 2f87939c-d615-4d13-8695-971e83ca7843] Refreshing network info cache for port 232b0dca-d02a-4796-921b-b5dca87b0a1b {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 715.222729] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-31c84044-b79b-4429-b5d3-d709ce79d827 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.234019] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ebf4bd3-19ce-4d6b-9ba0-8b4fd66ff440 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.258583] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] [instance: 2f87939c-d615-4d13-8695-971e83ca7843] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2f87939c-d615-4d13-8695-971e83ca7843 could not be found. [ 715.258583] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] [instance: 2f87939c-d615-4d13-8695-971e83ca7843] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 715.258678] env[63202]: INFO nova.compute.manager [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] [instance: 2f87939c-d615-4d13-8695-971e83ca7843] Took 0.04 seconds to destroy the instance on the hypervisor. [ 715.258893] env[63202]: DEBUG oslo.service.loopingcall [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 715.259136] env[63202]: DEBUG nova.compute.manager [-] [instance: 2f87939c-d615-4d13-8695-971e83ca7843] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 715.259227] env[63202]: DEBUG nova.network.neutron [-] [instance: 2f87939c-d615-4d13-8695-971e83ca7843] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 715.276106] env[63202]: DEBUG nova.network.neutron [-] [instance: 2f87939c-d615-4d13-8695-971e83ca7843] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 715.341848] env[63202]: DEBUG nova.network.neutron [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] Successfully created port: 501a66a6-4cf1-4ec6-8953-5eb6cbed571e {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 715.530711] env[63202]: DEBUG nova.compute.manager [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 715.747789] env[63202]: DEBUG nova.network.neutron [req-5c968337-ef92-4d43-94e4-6706189167c0 req-04c668f6-50e1-4d30-b924-f5f837a26d78 service nova] [instance: 2f87939c-d615-4d13-8695-971e83ca7843] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 715.780825] env[63202]: DEBUG nova.network.neutron [-] [instance: 2f87939c-d615-4d13-8695-971e83ca7843] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.893688] env[63202]: DEBUG nova.network.neutron [req-5c968337-ef92-4d43-94e4-6706189167c0 req-04c668f6-50e1-4d30-b924-f5f837a26d78 service nova] [instance: 2f87939c-d615-4d13-8695-971e83ca7843] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.018120] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e108ac68-b023-49af-ba21-03926e215829 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.026981] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-643d539f-59e6-427a-b9f1-4905c3cf4c05 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.069273] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dd55dcd-fa81-4b55-ba24-e1e5107e70b1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.077311] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50b660d4-b594-4441-ae8d-5431f648e18f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.092615] env[63202]: DEBUG nova.compute.provider_tree [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 716.250579] env[63202]: DEBUG nova.compute.manager [req-f98173e6-fd28-4494-8e66-7fde1ceeaa1d req-4c10434f-ed42-48d8-80c7-50bf4d7bc56e service nova] [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] Received event network-changed-501a66a6-4cf1-4ec6-8953-5eb6cbed571e {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 716.250804] env[63202]: DEBUG nova.compute.manager [req-f98173e6-fd28-4494-8e66-7fde1ceeaa1d req-4c10434f-ed42-48d8-80c7-50bf4d7bc56e service nova] [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] Refreshing instance network info cache due to event network-changed-501a66a6-4cf1-4ec6-8953-5eb6cbed571e. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 716.251025] env[63202]: DEBUG oslo_concurrency.lockutils [req-f98173e6-fd28-4494-8e66-7fde1ceeaa1d req-4c10434f-ed42-48d8-80c7-50bf4d7bc56e service nova] Acquiring lock "refresh_cache-8498aa14-fc1a-42e4-a5e6-e22b239a8f7e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.251205] env[63202]: DEBUG oslo_concurrency.lockutils [req-f98173e6-fd28-4494-8e66-7fde1ceeaa1d req-4c10434f-ed42-48d8-80c7-50bf4d7bc56e service nova] Acquired lock "refresh_cache-8498aa14-fc1a-42e4-a5e6-e22b239a8f7e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.251318] env[63202]: DEBUG nova.network.neutron [req-f98173e6-fd28-4494-8e66-7fde1ceeaa1d req-4c10434f-ed42-48d8-80c7-50bf4d7bc56e service nova] [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] Refreshing network info cache for port 501a66a6-4cf1-4ec6-8953-5eb6cbed571e {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 716.287320] env[63202]: INFO nova.compute.manager [-] [instance: 2f87939c-d615-4d13-8695-971e83ca7843] Took 1.03 seconds to deallocate network for instance. [ 716.289604] env[63202]: DEBUG nova.compute.claims [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] [instance: 2f87939c-d615-4d13-8695-971e83ca7843] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 716.289776] env[63202]: DEBUG oslo_concurrency.lockutils [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.318403] env[63202]: ERROR nova.compute.manager [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 501a66a6-4cf1-4ec6-8953-5eb6cbed571e, please check neutron logs for more information. [ 716.318403] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 716.318403] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 716.318403] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 716.318403] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 716.318403] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 716.318403] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 716.318403] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 716.318403] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 716.318403] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 716.318403] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 716.318403] env[63202]: ERROR nova.compute.manager raise self.value [ 716.318403] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 716.318403] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 716.318403] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 716.318403] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 716.318909] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 716.318909] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 716.318909] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 501a66a6-4cf1-4ec6-8953-5eb6cbed571e, please check neutron logs for more information. [ 716.318909] env[63202]: ERROR nova.compute.manager [ 716.318909] env[63202]: Traceback (most recent call last): [ 716.318909] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 716.318909] env[63202]: listener.cb(fileno) [ 716.318909] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 716.318909] env[63202]: result = function(*args, **kwargs) [ 716.318909] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 716.318909] env[63202]: return func(*args, **kwargs) [ 716.318909] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 716.318909] env[63202]: raise e [ 716.318909] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 716.318909] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 716.318909] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 716.318909] env[63202]: created_port_ids = self._update_ports_for_instance( [ 716.318909] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 716.318909] env[63202]: with excutils.save_and_reraise_exception(): [ 716.318909] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 716.318909] env[63202]: self.force_reraise() [ 716.318909] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 716.318909] env[63202]: raise self.value [ 716.318909] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 716.318909] env[63202]: updated_port = self._update_port( [ 716.318909] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 716.318909] env[63202]: _ensure_no_port_binding_failure(port) [ 716.318909] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 716.318909] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 716.319742] env[63202]: nova.exception.PortBindingFailed: Binding failed for port 501a66a6-4cf1-4ec6-8953-5eb6cbed571e, please check neutron logs for more information. [ 716.319742] env[63202]: Removing descriptor: 20 [ 716.397144] env[63202]: DEBUG oslo_concurrency.lockutils [req-5c968337-ef92-4d43-94e4-6706189167c0 req-04c668f6-50e1-4d30-b924-f5f837a26d78 service nova] Releasing lock "refresh_cache-2f87939c-d615-4d13-8695-971e83ca7843" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 716.397364] env[63202]: DEBUG nova.compute.manager [req-5c968337-ef92-4d43-94e4-6706189167c0 req-04c668f6-50e1-4d30-b924-f5f837a26d78 service nova] [instance: 2f87939c-d615-4d13-8695-971e83ca7843] Received event network-vif-deleted-232b0dca-d02a-4796-921b-b5dca87b0a1b {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 716.567972] env[63202]: DEBUG nova.compute.manager [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 716.593371] env[63202]: DEBUG nova.virt.hardware [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 716.593614] env[63202]: DEBUG nova.virt.hardware [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 716.593763] env[63202]: DEBUG nova.virt.hardware [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 716.593935] env[63202]: DEBUG nova.virt.hardware [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 716.594088] env[63202]: DEBUG nova.virt.hardware [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 716.594233] env[63202]: DEBUG nova.virt.hardware [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 716.594433] env[63202]: DEBUG nova.virt.hardware [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 716.594619] env[63202]: DEBUG nova.virt.hardware [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 716.594874] env[63202]: DEBUG nova.virt.hardware [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 716.595138] env[63202]: DEBUG nova.virt.hardware [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 716.595374] env[63202]: DEBUG nova.virt.hardware [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 716.596204] env[63202]: DEBUG nova.scheduler.client.report [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 716.600327] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0fe431d-da72-4992-a8d2-a484e68aaba1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.609326] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc0cec3c-c00b-4547-ad0b-043f2d49bd83 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.624375] env[63202]: ERROR nova.compute.manager [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 501a66a6-4cf1-4ec6-8953-5eb6cbed571e, please check neutron logs for more information. [ 716.624375] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] Traceback (most recent call last): [ 716.624375] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 716.624375] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] yield resources [ 716.624375] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 716.624375] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] self.driver.spawn(context, instance, image_meta, [ 716.624375] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 716.624375] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 716.624375] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 716.624375] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] vm_ref = self.build_virtual_machine(instance, [ 716.624375] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 716.624699] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] vif_infos = vmwarevif.get_vif_info(self._session, [ 716.624699] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 716.624699] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] for vif in network_info: [ 716.624699] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 716.624699] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] return self._sync_wrapper(fn, *args, **kwargs) [ 716.624699] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 716.624699] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] self.wait() [ 716.624699] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 716.624699] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] self[:] = self._gt.wait() [ 716.624699] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 716.624699] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] return self._exit_event.wait() [ 716.624699] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 716.624699] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] current.throw(*self._exc) [ 716.625039] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 716.625039] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] result = function(*args, **kwargs) [ 716.625039] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 716.625039] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] return func(*args, **kwargs) [ 716.625039] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 716.625039] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] raise e [ 716.625039] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 716.625039] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] nwinfo = self.network_api.allocate_for_instance( [ 716.625039] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 716.625039] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] created_port_ids = self._update_ports_for_instance( [ 716.625039] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 716.625039] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] with excutils.save_and_reraise_exception(): [ 716.625039] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 716.625620] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] self.force_reraise() [ 716.625620] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 716.625620] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] raise self.value [ 716.625620] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 716.625620] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] updated_port = self._update_port( [ 716.625620] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 716.625620] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] _ensure_no_port_binding_failure(port) [ 716.625620] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 716.625620] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] raise exception.PortBindingFailed(port_id=port['id']) [ 716.625620] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] nova.exception.PortBindingFailed: Binding failed for port 501a66a6-4cf1-4ec6-8953-5eb6cbed571e, please check neutron logs for more information. [ 716.625620] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] [ 716.625620] env[63202]: INFO nova.compute.manager [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] Terminating instance [ 716.627233] env[63202]: DEBUG oslo_concurrency.lockutils [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Acquiring lock "refresh_cache-8498aa14-fc1a-42e4-a5e6-e22b239a8f7e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.769416] env[63202]: DEBUG nova.network.neutron [req-f98173e6-fd28-4494-8e66-7fde1ceeaa1d req-4c10434f-ed42-48d8-80c7-50bf4d7bc56e service nova] [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 716.842996] env[63202]: DEBUG nova.network.neutron [req-f98173e6-fd28-4494-8e66-7fde1ceeaa1d req-4c10434f-ed42-48d8-80c7-50bf4d7bc56e service nova] [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.104617] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.589s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.105142] env[63202]: DEBUG nova.compute.manager [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 06496aad-b025-48c2-8436-03d43d1ac899] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 717.107907] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.482s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 717.109478] env[63202]: INFO nova.compute.claims [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 717.345626] env[63202]: DEBUG oslo_concurrency.lockutils [req-f98173e6-fd28-4494-8e66-7fde1ceeaa1d req-4c10434f-ed42-48d8-80c7-50bf4d7bc56e service nova] Releasing lock "refresh_cache-8498aa14-fc1a-42e4-a5e6-e22b239a8f7e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.346042] env[63202]: DEBUG oslo_concurrency.lockutils [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Acquired lock "refresh_cache-8498aa14-fc1a-42e4-a5e6-e22b239a8f7e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.346248] env[63202]: DEBUG nova.network.neutron [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 717.618106] env[63202]: DEBUG nova.compute.utils [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 717.618106] env[63202]: DEBUG nova.compute.manager [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 06496aad-b025-48c2-8436-03d43d1ac899] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 717.618106] env[63202]: DEBUG nova.network.neutron [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 06496aad-b025-48c2-8436-03d43d1ac899] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 717.656840] env[63202]: DEBUG nova.policy [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '24878273948f4c2584621b6eb3f79f5a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a8318d94cc5048de936eb06da4642545', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 717.862889] env[63202]: DEBUG nova.network.neutron [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 717.896012] env[63202]: DEBUG nova.network.neutron [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 06496aad-b025-48c2-8436-03d43d1ac899] Successfully created port: ce51d939-c2c1-4116-9d39-577687daaf4f {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 717.951889] env[63202]: DEBUG nova.network.neutron [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.122715] env[63202]: DEBUG nova.compute.manager [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 06496aad-b025-48c2-8436-03d43d1ac899] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 718.301597] env[63202]: DEBUG nova.compute.manager [req-84cc7fcf-28c8-4253-b71f-37f353e83145 req-93daf1ca-4b0a-4e6c-a764-f49bccabf098 service nova] [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] Received event network-vif-deleted-501a66a6-4cf1-4ec6-8953-5eb6cbed571e {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 718.454319] env[63202]: DEBUG oslo_concurrency.lockutils [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Releasing lock "refresh_cache-8498aa14-fc1a-42e4-a5e6-e22b239a8f7e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.454720] env[63202]: DEBUG nova.compute.manager [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 718.454900] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 718.455202] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-aefff613-daf8-4d0f-a94e-77d8e1c797e3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.466184] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17b7e314-4857-41f9-895e-2d77a7f7cb2b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.490599] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e could not be found. [ 718.490830] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 718.491015] env[63202]: INFO nova.compute.manager [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 718.491255] env[63202]: DEBUG oslo.service.loopingcall [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 718.493567] env[63202]: DEBUG nova.compute.manager [-] [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 718.493669] env[63202]: DEBUG nova.network.neutron [-] [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 718.512751] env[63202]: DEBUG nova.network.neutron [-] [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 718.604430] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b3844d2-9a64-4c4b-965d-1e3ff0bb787d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.612227] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77fa8b8f-0023-496c-b835-66b3f925d03b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.649676] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-145fa9f9-6aa8-40d1-984c-3e5b30523db6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.657446] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dba6bec2-1124-480d-83e1-b43ff4a2a0ec {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.670623] env[63202]: DEBUG nova.compute.provider_tree [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 718.698028] env[63202]: ERROR nova.compute.manager [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ce51d939-c2c1-4116-9d39-577687daaf4f, please check neutron logs for more information. [ 718.698028] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 718.698028] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 718.698028] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 718.698028] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 718.698028] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 718.698028] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 718.698028] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 718.698028] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.698028] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 718.698028] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.698028] env[63202]: ERROR nova.compute.manager raise self.value [ 718.698028] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 718.698028] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 718.698028] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.698028] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 718.699227] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.699227] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 718.699227] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ce51d939-c2c1-4116-9d39-577687daaf4f, please check neutron logs for more information. [ 718.699227] env[63202]: ERROR nova.compute.manager [ 718.699227] env[63202]: Traceback (most recent call last): [ 718.699227] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 718.699227] env[63202]: listener.cb(fileno) [ 718.699227] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 718.699227] env[63202]: result = function(*args, **kwargs) [ 718.699227] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 718.699227] env[63202]: return func(*args, **kwargs) [ 718.699227] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 718.699227] env[63202]: raise e [ 718.699227] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 718.699227] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 718.699227] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 718.699227] env[63202]: created_port_ids = self._update_ports_for_instance( [ 718.699227] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 718.699227] env[63202]: with excutils.save_and_reraise_exception(): [ 718.699227] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.699227] env[63202]: self.force_reraise() [ 718.699227] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.699227] env[63202]: raise self.value [ 718.699227] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 718.699227] env[63202]: updated_port = self._update_port( [ 718.699227] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.699227] env[63202]: _ensure_no_port_binding_failure(port) [ 718.699227] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.699227] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 718.699995] env[63202]: nova.exception.PortBindingFailed: Binding failed for port ce51d939-c2c1-4116-9d39-577687daaf4f, please check neutron logs for more information. [ 718.699995] env[63202]: Removing descriptor: 20 [ 719.016613] env[63202]: DEBUG nova.network.neutron [-] [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.154322] env[63202]: DEBUG nova.compute.manager [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 06496aad-b025-48c2-8436-03d43d1ac899] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 719.173632] env[63202]: DEBUG nova.scheduler.client.report [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 719.182598] env[63202]: DEBUG nova.virt.hardware [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 719.182915] env[63202]: DEBUG nova.virt.hardware [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 719.183096] env[63202]: DEBUG nova.virt.hardware [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 719.183276] env[63202]: DEBUG nova.virt.hardware [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 719.183411] env[63202]: DEBUG nova.virt.hardware [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 719.184365] env[63202]: DEBUG nova.virt.hardware [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 719.184365] env[63202]: DEBUG nova.virt.hardware [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 719.184365] env[63202]: DEBUG nova.virt.hardware [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 719.184365] env[63202]: DEBUG nova.virt.hardware [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 719.184365] env[63202]: DEBUG nova.virt.hardware [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 719.184537] env[63202]: DEBUG nova.virt.hardware [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 719.186154] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84e1a728-dfc3-4410-922b-5f0233049ae9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.194169] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b871411c-da15-4311-a856-60d5caa4c080 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.208646] env[63202]: ERROR nova.compute.manager [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 06496aad-b025-48c2-8436-03d43d1ac899] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ce51d939-c2c1-4116-9d39-577687daaf4f, please check neutron logs for more information. [ 719.208646] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] Traceback (most recent call last): [ 719.208646] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 719.208646] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] yield resources [ 719.208646] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 719.208646] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] self.driver.spawn(context, instance, image_meta, [ 719.208646] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 719.208646] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] self._vmops.spawn(context, instance, image_meta, injected_files, [ 719.208646] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 719.208646] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] vm_ref = self.build_virtual_machine(instance, [ 719.208646] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 719.209075] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] vif_infos = vmwarevif.get_vif_info(self._session, [ 719.209075] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 719.209075] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] for vif in network_info: [ 719.209075] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 719.209075] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] return self._sync_wrapper(fn, *args, **kwargs) [ 719.209075] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 719.209075] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] self.wait() [ 719.209075] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 719.209075] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] self[:] = self._gt.wait() [ 719.209075] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 719.209075] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] return self._exit_event.wait() [ 719.209075] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 719.209075] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] current.throw(*self._exc) [ 719.209451] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 719.209451] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] result = function(*args, **kwargs) [ 719.209451] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 719.209451] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] return func(*args, **kwargs) [ 719.209451] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 719.209451] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] raise e [ 719.209451] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 719.209451] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] nwinfo = self.network_api.allocate_for_instance( [ 719.209451] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 719.209451] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] created_port_ids = self._update_ports_for_instance( [ 719.209451] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 719.209451] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] with excutils.save_and_reraise_exception(): [ 719.209451] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 719.209865] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] self.force_reraise() [ 719.209865] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 719.209865] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] raise self.value [ 719.209865] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 719.209865] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] updated_port = self._update_port( [ 719.209865] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 719.209865] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] _ensure_no_port_binding_failure(port) [ 719.209865] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 719.209865] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] raise exception.PortBindingFailed(port_id=port['id']) [ 719.209865] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] nova.exception.PortBindingFailed: Binding failed for port ce51d939-c2c1-4116-9d39-577687daaf4f, please check neutron logs for more information. [ 719.209865] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] [ 719.209865] env[63202]: INFO nova.compute.manager [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 06496aad-b025-48c2-8436-03d43d1ac899] Terminating instance [ 719.210813] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Acquiring lock "refresh_cache-06496aad-b025-48c2-8436-03d43d1ac899" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.210963] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Acquired lock "refresh_cache-06496aad-b025-48c2-8436-03d43d1ac899" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.211148] env[63202]: DEBUG nova.network.neutron [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 06496aad-b025-48c2-8436-03d43d1ac899] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 719.521578] env[63202]: INFO nova.compute.manager [-] [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] Took 1.03 seconds to deallocate network for instance. [ 719.524049] env[63202]: DEBUG nova.compute.claims [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 719.524142] env[63202]: DEBUG oslo_concurrency.lockutils [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 719.678168] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.570s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 719.678776] env[63202]: DEBUG nova.compute.manager [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 719.681502] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.993s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 719.683190] env[63202]: INFO nova.compute.claims [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] [instance: a97cff65-e350-480e-9891-3317fe05be47] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 719.729095] env[63202]: DEBUG nova.network.neutron [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 06496aad-b025-48c2-8436-03d43d1ac899] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 719.820964] env[63202]: DEBUG nova.network.neutron [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 06496aad-b025-48c2-8436-03d43d1ac899] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.187566] env[63202]: DEBUG nova.compute.utils [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 720.190984] env[63202]: DEBUG nova.compute.manager [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 720.191163] env[63202]: DEBUG nova.network.neutron [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 720.229293] env[63202]: DEBUG nova.policy [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '569158eee09b4bd597a32db0c2f02948', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '43717c26a67e4ea7bc9439ef2475b794', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 720.326142] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Releasing lock "refresh_cache-06496aad-b025-48c2-8436-03d43d1ac899" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.326142] env[63202]: DEBUG nova.compute.manager [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 06496aad-b025-48c2-8436-03d43d1ac899] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 720.326142] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 06496aad-b025-48c2-8436-03d43d1ac899] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 720.327010] env[63202]: DEBUG nova.compute.manager [req-0fa3f29b-d422-46e4-92c6-ebd23a17a0b9 req-4d16f26a-eb8f-4117-b7e0-a31d75b17228 service nova] [instance: 06496aad-b025-48c2-8436-03d43d1ac899] Received event network-changed-ce51d939-c2c1-4116-9d39-577687daaf4f {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 720.327194] env[63202]: DEBUG nova.compute.manager [req-0fa3f29b-d422-46e4-92c6-ebd23a17a0b9 req-4d16f26a-eb8f-4117-b7e0-a31d75b17228 service nova] [instance: 06496aad-b025-48c2-8436-03d43d1ac899] Refreshing instance network info cache due to event network-changed-ce51d939-c2c1-4116-9d39-577687daaf4f. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 720.327397] env[63202]: DEBUG oslo_concurrency.lockutils [req-0fa3f29b-d422-46e4-92c6-ebd23a17a0b9 req-4d16f26a-eb8f-4117-b7e0-a31d75b17228 service nova] Acquiring lock "refresh_cache-06496aad-b025-48c2-8436-03d43d1ac899" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 720.327533] env[63202]: DEBUG oslo_concurrency.lockutils [req-0fa3f29b-d422-46e4-92c6-ebd23a17a0b9 req-4d16f26a-eb8f-4117-b7e0-a31d75b17228 service nova] Acquired lock "refresh_cache-06496aad-b025-48c2-8436-03d43d1ac899" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.327717] env[63202]: DEBUG nova.network.neutron [req-0fa3f29b-d422-46e4-92c6-ebd23a17a0b9 req-4d16f26a-eb8f-4117-b7e0-a31d75b17228 service nova] [instance: 06496aad-b025-48c2-8436-03d43d1ac899] Refreshing network info cache for port ce51d939-c2c1-4116-9d39-577687daaf4f {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 720.329019] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-84c33be5-37e2-4120-9c96-67c812d716f1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.339872] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6aa3162a-7e0c-444b-bdc8-4e02b09f352b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.362165] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 06496aad-b025-48c2-8436-03d43d1ac899] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 06496aad-b025-48c2-8436-03d43d1ac899 could not be found. [ 720.362394] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 06496aad-b025-48c2-8436-03d43d1ac899] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 720.362566] env[63202]: INFO nova.compute.manager [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 06496aad-b025-48c2-8436-03d43d1ac899] Took 0.04 seconds to destroy the instance on the hypervisor. [ 720.362795] env[63202]: DEBUG oslo.service.loopingcall [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 720.363022] env[63202]: DEBUG nova.compute.manager [-] [instance: 06496aad-b025-48c2-8436-03d43d1ac899] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 720.363110] env[63202]: DEBUG nova.network.neutron [-] [instance: 06496aad-b025-48c2-8436-03d43d1ac899] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 720.379159] env[63202]: DEBUG nova.network.neutron [-] [instance: 06496aad-b025-48c2-8436-03d43d1ac899] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 720.513741] env[63202]: DEBUG nova.network.neutron [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] Successfully created port: 09ee3acd-6615-4372-b295-b69bb039fbce {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 720.691732] env[63202]: DEBUG nova.compute.manager [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 720.867988] env[63202]: DEBUG nova.network.neutron [req-0fa3f29b-d422-46e4-92c6-ebd23a17a0b9 req-4d16f26a-eb8f-4117-b7e0-a31d75b17228 service nova] [instance: 06496aad-b025-48c2-8436-03d43d1ac899] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 720.882384] env[63202]: DEBUG nova.network.neutron [-] [instance: 06496aad-b025-48c2-8436-03d43d1ac899] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.973443] env[63202]: DEBUG nova.network.neutron [req-0fa3f29b-d422-46e4-92c6-ebd23a17a0b9 req-4d16f26a-eb8f-4117-b7e0-a31d75b17228 service nova] [instance: 06496aad-b025-48c2-8436-03d43d1ac899] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.096677] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a83c7ff9-a1e4-4f4f-a21f-5646b0be93e0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.105325] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd648019-e512-4576-9727-6af903a88e49 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.139756] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3d811af-7628-4399-9537-56f44f5b41f3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.147291] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf6755b0-8c5e-482e-bc09-918a4804955e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.160629] env[63202]: DEBUG nova.compute.provider_tree [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 721.353323] env[63202]: ERROR nova.compute.manager [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 09ee3acd-6615-4372-b295-b69bb039fbce, please check neutron logs for more information. [ 721.353323] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 721.353323] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 721.353323] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 721.353323] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 721.353323] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 721.353323] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 721.353323] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 721.353323] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 721.353323] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 721.353323] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 721.353323] env[63202]: ERROR nova.compute.manager raise self.value [ 721.353323] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 721.353323] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 721.353323] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 721.353323] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 721.353978] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 721.353978] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 721.353978] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 09ee3acd-6615-4372-b295-b69bb039fbce, please check neutron logs for more information. [ 721.353978] env[63202]: ERROR nova.compute.manager [ 721.353978] env[63202]: Traceback (most recent call last): [ 721.353978] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 721.353978] env[63202]: listener.cb(fileno) [ 721.353978] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 721.353978] env[63202]: result = function(*args, **kwargs) [ 721.353978] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 721.353978] env[63202]: return func(*args, **kwargs) [ 721.353978] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 721.353978] env[63202]: raise e [ 721.353978] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 721.353978] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 721.353978] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 721.353978] env[63202]: created_port_ids = self._update_ports_for_instance( [ 721.353978] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 721.353978] env[63202]: with excutils.save_and_reraise_exception(): [ 721.353978] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 721.353978] env[63202]: self.force_reraise() [ 721.353978] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 721.353978] env[63202]: raise self.value [ 721.353978] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 721.353978] env[63202]: updated_port = self._update_port( [ 721.353978] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 721.353978] env[63202]: _ensure_no_port_binding_failure(port) [ 721.353978] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 721.353978] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 721.354808] env[63202]: nova.exception.PortBindingFailed: Binding failed for port 09ee3acd-6615-4372-b295-b69bb039fbce, please check neutron logs for more information. [ 721.354808] env[63202]: Removing descriptor: 20 [ 721.385410] env[63202]: INFO nova.compute.manager [-] [instance: 06496aad-b025-48c2-8436-03d43d1ac899] Took 1.02 seconds to deallocate network for instance. [ 721.387705] env[63202]: DEBUG nova.compute.claims [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 06496aad-b025-48c2-8436-03d43d1ac899] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 721.387882] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 721.475657] env[63202]: DEBUG oslo_concurrency.lockutils [req-0fa3f29b-d422-46e4-92c6-ebd23a17a0b9 req-4d16f26a-eb8f-4117-b7e0-a31d75b17228 service nova] Releasing lock "refresh_cache-06496aad-b025-48c2-8436-03d43d1ac899" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 721.475912] env[63202]: DEBUG nova.compute.manager [req-0fa3f29b-d422-46e4-92c6-ebd23a17a0b9 req-4d16f26a-eb8f-4117-b7e0-a31d75b17228 service nova] [instance: 06496aad-b025-48c2-8436-03d43d1ac899] Received event network-vif-deleted-ce51d939-c2c1-4116-9d39-577687daaf4f {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 721.663629] env[63202]: DEBUG nova.scheduler.client.report [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 721.704933] env[63202]: DEBUG nova.compute.manager [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 721.732612] env[63202]: DEBUG nova.virt.hardware [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 721.732857] env[63202]: DEBUG nova.virt.hardware [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 721.733034] env[63202]: DEBUG nova.virt.hardware [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 721.733224] env[63202]: DEBUG nova.virt.hardware [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 721.733369] env[63202]: DEBUG nova.virt.hardware [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 721.733575] env[63202]: DEBUG nova.virt.hardware [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 721.733794] env[63202]: DEBUG nova.virt.hardware [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 721.734455] env[63202]: DEBUG nova.virt.hardware [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 721.734455] env[63202]: DEBUG nova.virt.hardware [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 721.734455] env[63202]: DEBUG nova.virt.hardware [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 721.734455] env[63202]: DEBUG nova.virt.hardware [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 721.735775] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba0a9472-9455-41b1-a917-3d1650271ee9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.743022] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0323aa1c-2795-46b8-99ae-270a83c837bb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.756651] env[63202]: ERROR nova.compute.manager [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 09ee3acd-6615-4372-b295-b69bb039fbce, please check neutron logs for more information. [ 721.756651] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] Traceback (most recent call last): [ 721.756651] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 721.756651] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] yield resources [ 721.756651] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 721.756651] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] self.driver.spawn(context, instance, image_meta, [ 721.756651] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 721.756651] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] self._vmops.spawn(context, instance, image_meta, injected_files, [ 721.756651] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 721.756651] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] vm_ref = self.build_virtual_machine(instance, [ 721.756651] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 721.757035] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] vif_infos = vmwarevif.get_vif_info(self._session, [ 721.757035] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 721.757035] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] for vif in network_info: [ 721.757035] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 721.757035] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] return self._sync_wrapper(fn, *args, **kwargs) [ 721.757035] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 721.757035] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] self.wait() [ 721.757035] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 721.757035] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] self[:] = self._gt.wait() [ 721.757035] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 721.757035] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] return self._exit_event.wait() [ 721.757035] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 721.757035] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] current.throw(*self._exc) [ 721.757424] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 721.757424] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] result = function(*args, **kwargs) [ 721.757424] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 721.757424] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] return func(*args, **kwargs) [ 721.757424] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 721.757424] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] raise e [ 721.757424] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 721.757424] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] nwinfo = self.network_api.allocate_for_instance( [ 721.757424] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 721.757424] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] created_port_ids = self._update_ports_for_instance( [ 721.757424] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 721.757424] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] with excutils.save_and_reraise_exception(): [ 721.757424] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 721.757801] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] self.force_reraise() [ 721.757801] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 721.757801] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] raise self.value [ 721.757801] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 721.757801] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] updated_port = self._update_port( [ 721.757801] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 721.757801] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] _ensure_no_port_binding_failure(port) [ 721.757801] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 721.757801] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] raise exception.PortBindingFailed(port_id=port['id']) [ 721.757801] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] nova.exception.PortBindingFailed: Binding failed for port 09ee3acd-6615-4372-b295-b69bb039fbce, please check neutron logs for more information. [ 721.757801] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] [ 721.757801] env[63202]: INFO nova.compute.manager [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] Terminating instance [ 721.758914] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Acquiring lock "refresh_cache-8885156e-78f1-45ff-87c9-829f3ab89dca" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 721.759076] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Acquired lock "refresh_cache-8885156e-78f1-45ff-87c9-829f3ab89dca" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.759239] env[63202]: DEBUG nova.network.neutron [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 722.168768] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.486s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.168768] env[63202]: DEBUG nova.compute.manager [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] [instance: a97cff65-e350-480e-9891-3317fe05be47] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 722.171419] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 35.063s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 722.171593] env[63202]: DEBUG nova.objects.instance [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63202) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 722.277360] env[63202]: DEBUG nova.network.neutron [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 722.352213] env[63202]: DEBUG nova.compute.manager [req-ad9c2aa9-5531-4290-9339-95b5271de2d0 req-24ced1d9-29f8-4bce-91e1-a3600f0940aa service nova] [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] Received event network-changed-09ee3acd-6615-4372-b295-b69bb039fbce {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 722.352378] env[63202]: DEBUG nova.compute.manager [req-ad9c2aa9-5531-4290-9339-95b5271de2d0 req-24ced1d9-29f8-4bce-91e1-a3600f0940aa service nova] [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] Refreshing instance network info cache due to event network-changed-09ee3acd-6615-4372-b295-b69bb039fbce. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 722.352479] env[63202]: DEBUG oslo_concurrency.lockutils [req-ad9c2aa9-5531-4290-9339-95b5271de2d0 req-24ced1d9-29f8-4bce-91e1-a3600f0940aa service nova] Acquiring lock "refresh_cache-8885156e-78f1-45ff-87c9-829f3ab89dca" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 722.370163] env[63202]: DEBUG nova.network.neutron [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.676231] env[63202]: DEBUG nova.compute.utils [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 722.681227] env[63202]: DEBUG nova.compute.manager [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] [instance: a97cff65-e350-480e-9891-3317fe05be47] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 722.681454] env[63202]: DEBUG nova.network.neutron [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] [instance: a97cff65-e350-480e-9891-3317fe05be47] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 722.719835] env[63202]: DEBUG nova.policy [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f64d1aee5efa45e68916b92777a359ef', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3b4afd7bfb434a79bd251dbc986b12c4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 722.872602] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Releasing lock "refresh_cache-8885156e-78f1-45ff-87c9-829f3ab89dca" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 722.873016] env[63202]: DEBUG nova.compute.manager [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 722.873234] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 722.873531] env[63202]: DEBUG oslo_concurrency.lockutils [req-ad9c2aa9-5531-4290-9339-95b5271de2d0 req-24ced1d9-29f8-4bce-91e1-a3600f0940aa service nova] Acquired lock "refresh_cache-8885156e-78f1-45ff-87c9-829f3ab89dca" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 722.873698] env[63202]: DEBUG nova.network.neutron [req-ad9c2aa9-5531-4290-9339-95b5271de2d0 req-24ced1d9-29f8-4bce-91e1-a3600f0940aa service nova] [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] Refreshing network info cache for port 09ee3acd-6615-4372-b295-b69bb039fbce {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 722.875242] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0047da7a-0ae9-48b2-9885-2b8b8cb35b6c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.885096] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c3864a3-674f-44c4-9569-576b1bb1baa8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.908382] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8885156e-78f1-45ff-87c9-829f3ab89dca could not be found. [ 722.908603] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 722.908787] env[63202]: INFO nova.compute.manager [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] Took 0.04 seconds to destroy the instance on the hypervisor. [ 722.909032] env[63202]: DEBUG oslo.service.loopingcall [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 722.909245] env[63202]: DEBUG nova.compute.manager [-] [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 722.909335] env[63202]: DEBUG nova.network.neutron [-] [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 722.922752] env[63202]: DEBUG nova.network.neutron [-] [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 722.962636] env[63202]: DEBUG nova.network.neutron [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] [instance: a97cff65-e350-480e-9891-3317fe05be47] Successfully created port: d12cd099-0585-4a87-9039-680bda744ef9 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 723.185225] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e6877d10-1cbb-4536-b77e-a71d76ccea22 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.189468] env[63202]: DEBUG nova.compute.manager [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] [instance: a97cff65-e350-480e-9891-3317fe05be47] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 723.190845] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 34.416s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.192272] env[63202]: INFO nova.compute.claims [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 723.411807] env[63202]: DEBUG nova.network.neutron [req-ad9c2aa9-5531-4290-9339-95b5271de2d0 req-24ced1d9-29f8-4bce-91e1-a3600f0940aa service nova] [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 723.425247] env[63202]: DEBUG nova.network.neutron [-] [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.524479] env[63202]: DEBUG nova.network.neutron [req-ad9c2aa9-5531-4290-9339-95b5271de2d0 req-24ced1d9-29f8-4bce-91e1-a3600f0940aa service nova] [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.812214] env[63202]: ERROR nova.compute.manager [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d12cd099-0585-4a87-9039-680bda744ef9, please check neutron logs for more information. [ 723.812214] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 723.812214] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 723.812214] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 723.812214] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 723.812214] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 723.812214] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 723.812214] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 723.812214] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 723.812214] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 723.812214] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 723.812214] env[63202]: ERROR nova.compute.manager raise self.value [ 723.812214] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 723.812214] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 723.812214] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 723.812214] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 723.812698] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 723.812698] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 723.812698] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d12cd099-0585-4a87-9039-680bda744ef9, please check neutron logs for more information. [ 723.812698] env[63202]: ERROR nova.compute.manager [ 723.812698] env[63202]: Traceback (most recent call last): [ 723.812698] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 723.812698] env[63202]: listener.cb(fileno) [ 723.812698] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 723.812698] env[63202]: result = function(*args, **kwargs) [ 723.812698] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 723.812698] env[63202]: return func(*args, **kwargs) [ 723.812698] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 723.812698] env[63202]: raise e [ 723.812698] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 723.812698] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 723.812698] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 723.812698] env[63202]: created_port_ids = self._update_ports_for_instance( [ 723.812698] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 723.812698] env[63202]: with excutils.save_and_reraise_exception(): [ 723.812698] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 723.812698] env[63202]: self.force_reraise() [ 723.812698] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 723.812698] env[63202]: raise self.value [ 723.812698] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 723.812698] env[63202]: updated_port = self._update_port( [ 723.812698] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 723.812698] env[63202]: _ensure_no_port_binding_failure(port) [ 723.812698] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 723.812698] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 723.813513] env[63202]: nova.exception.PortBindingFailed: Binding failed for port d12cd099-0585-4a87-9039-680bda744ef9, please check neutron logs for more information. [ 723.813513] env[63202]: Removing descriptor: 20 [ 723.927913] env[63202]: INFO nova.compute.manager [-] [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] Took 1.02 seconds to deallocate network for instance. [ 723.930747] env[63202]: DEBUG nova.compute.claims [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 723.930922] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.027633] env[63202]: DEBUG oslo_concurrency.lockutils [req-ad9c2aa9-5531-4290-9339-95b5271de2d0 req-24ced1d9-29f8-4bce-91e1-a3600f0940aa service nova] Releasing lock "refresh_cache-8885156e-78f1-45ff-87c9-829f3ab89dca" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 724.027812] env[63202]: DEBUG nova.compute.manager [req-ad9c2aa9-5531-4290-9339-95b5271de2d0 req-24ced1d9-29f8-4bce-91e1-a3600f0940aa service nova] [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] Received event network-vif-deleted-09ee3acd-6615-4372-b295-b69bb039fbce {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 724.196414] env[63202]: DEBUG nova.compute.manager [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] [instance: a97cff65-e350-480e-9891-3317fe05be47] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 724.221958] env[63202]: DEBUG nova.virt.hardware [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 724.222273] env[63202]: DEBUG nova.virt.hardware [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 724.222382] env[63202]: DEBUG nova.virt.hardware [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 724.222560] env[63202]: DEBUG nova.virt.hardware [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 724.222705] env[63202]: DEBUG nova.virt.hardware [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 724.222892] env[63202]: DEBUG nova.virt.hardware [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 724.223112] env[63202]: DEBUG nova.virt.hardware [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 724.223282] env[63202]: DEBUG nova.virt.hardware [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 724.223432] env[63202]: DEBUG nova.virt.hardware [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 724.223592] env[63202]: DEBUG nova.virt.hardware [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 724.223762] env[63202]: DEBUG nova.virt.hardware [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 724.224688] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8669eb97-e52d-491c-b73c-e39d5080cdde {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.238025] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8690e0f-8002-4c2e-a9a3-ebce03d99764 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.250272] env[63202]: ERROR nova.compute.manager [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] [instance: a97cff65-e350-480e-9891-3317fe05be47] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d12cd099-0585-4a87-9039-680bda744ef9, please check neutron logs for more information. [ 724.250272] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] Traceback (most recent call last): [ 724.250272] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 724.250272] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] yield resources [ 724.250272] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 724.250272] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] self.driver.spawn(context, instance, image_meta, [ 724.250272] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 724.250272] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] self._vmops.spawn(context, instance, image_meta, injected_files, [ 724.250272] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 724.250272] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] vm_ref = self.build_virtual_machine(instance, [ 724.250272] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 724.250577] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] vif_infos = vmwarevif.get_vif_info(self._session, [ 724.250577] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 724.250577] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] for vif in network_info: [ 724.250577] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 724.250577] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] return self._sync_wrapper(fn, *args, **kwargs) [ 724.250577] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 724.250577] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] self.wait() [ 724.250577] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 724.250577] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] self[:] = self._gt.wait() [ 724.250577] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 724.250577] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] return self._exit_event.wait() [ 724.250577] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 724.250577] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] current.throw(*self._exc) [ 724.250895] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 724.250895] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] result = function(*args, **kwargs) [ 724.250895] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 724.250895] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] return func(*args, **kwargs) [ 724.250895] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 724.250895] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] raise e [ 724.250895] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 724.250895] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] nwinfo = self.network_api.allocate_for_instance( [ 724.250895] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 724.250895] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] created_port_ids = self._update_ports_for_instance( [ 724.250895] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 724.250895] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] with excutils.save_and_reraise_exception(): [ 724.250895] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 724.251210] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] self.force_reraise() [ 724.251210] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 724.251210] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] raise self.value [ 724.251210] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 724.251210] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] updated_port = self._update_port( [ 724.251210] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 724.251210] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] _ensure_no_port_binding_failure(port) [ 724.251210] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 724.251210] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] raise exception.PortBindingFailed(port_id=port['id']) [ 724.251210] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] nova.exception.PortBindingFailed: Binding failed for port d12cd099-0585-4a87-9039-680bda744ef9, please check neutron logs for more information. [ 724.251210] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] [ 724.251210] env[63202]: INFO nova.compute.manager [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] [instance: a97cff65-e350-480e-9891-3317fe05be47] Terminating instance [ 724.255100] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] Acquiring lock "refresh_cache-a97cff65-e350-480e-9891-3317fe05be47" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 724.255100] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] Acquired lock "refresh_cache-a97cff65-e350-480e-9891-3317fe05be47" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.255307] env[63202]: DEBUG nova.network.neutron [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] [instance: a97cff65-e350-480e-9891-3317fe05be47] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 724.532051] env[63202]: DEBUG nova.compute.manager [req-1eb02d81-2104-42e1-b618-55b8116b2baf req-d76b9ad4-fe47-445d-99b8-21283da72168 service nova] [instance: a97cff65-e350-480e-9891-3317fe05be47] Received event network-changed-d12cd099-0585-4a87-9039-680bda744ef9 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 724.532343] env[63202]: DEBUG nova.compute.manager [req-1eb02d81-2104-42e1-b618-55b8116b2baf req-d76b9ad4-fe47-445d-99b8-21283da72168 service nova] [instance: a97cff65-e350-480e-9891-3317fe05be47] Refreshing instance network info cache due to event network-changed-d12cd099-0585-4a87-9039-680bda744ef9. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 724.533063] env[63202]: DEBUG oslo_concurrency.lockutils [req-1eb02d81-2104-42e1-b618-55b8116b2baf req-d76b9ad4-fe47-445d-99b8-21283da72168 service nova] Acquiring lock "refresh_cache-a97cff65-e350-480e-9891-3317fe05be47" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 724.580119] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b44f99e9-ebe2-4d16-97a8-4c0d2eb24525 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.587622] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ccf61b6-71ed-443a-ab9c-d985a030cbcc {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.619995] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-133197e5-49f1-4c77-9ad1-fcb6a6ef1cb0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.627652] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-600f422f-048f-43a5-90cf-a59410ee8877 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.641331] env[63202]: DEBUG nova.compute.provider_tree [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 724.776739] env[63202]: DEBUG nova.network.neutron [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] [instance: a97cff65-e350-480e-9891-3317fe05be47] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 724.862069] env[63202]: DEBUG nova.network.neutron [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] [instance: a97cff65-e350-480e-9891-3317fe05be47] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.146713] env[63202]: DEBUG nova.scheduler.client.report [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 725.364892] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] Releasing lock "refresh_cache-a97cff65-e350-480e-9891-3317fe05be47" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.364892] env[63202]: DEBUG nova.compute.manager [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] [instance: a97cff65-e350-480e-9891-3317fe05be47] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 725.365110] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] [instance: a97cff65-e350-480e-9891-3317fe05be47] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 725.365427] env[63202]: DEBUG oslo_concurrency.lockutils [req-1eb02d81-2104-42e1-b618-55b8116b2baf req-d76b9ad4-fe47-445d-99b8-21283da72168 service nova] Acquired lock "refresh_cache-a97cff65-e350-480e-9891-3317fe05be47" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.365598] env[63202]: DEBUG nova.network.neutron [req-1eb02d81-2104-42e1-b618-55b8116b2baf req-d76b9ad4-fe47-445d-99b8-21283da72168 service nova] [instance: a97cff65-e350-480e-9891-3317fe05be47] Refreshing network info cache for port d12cd099-0585-4a87-9039-680bda744ef9 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 725.366664] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7cb392d5-515e-4409-85d1-a7425e6d5f45 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.376973] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbf1e1df-66ec-435f-9bbb-c412b351c887 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.399011] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] [instance: a97cff65-e350-480e-9891-3317fe05be47] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a97cff65-e350-480e-9891-3317fe05be47 could not be found. [ 725.399216] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] [instance: a97cff65-e350-480e-9891-3317fe05be47] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 725.399410] env[63202]: INFO nova.compute.manager [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] [instance: a97cff65-e350-480e-9891-3317fe05be47] Took 0.03 seconds to destroy the instance on the hypervisor. [ 725.399675] env[63202]: DEBUG oslo.service.loopingcall [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 725.399889] env[63202]: DEBUG nova.compute.manager [-] [instance: a97cff65-e350-480e-9891-3317fe05be47] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 725.399979] env[63202]: DEBUG nova.network.neutron [-] [instance: a97cff65-e350-480e-9891-3317fe05be47] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 725.414730] env[63202]: DEBUG nova.network.neutron [-] [instance: a97cff65-e350-480e-9891-3317fe05be47] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 725.651686] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.461s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.652336] env[63202]: DEBUG nova.compute.manager [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 725.655162] env[63202]: DEBUG oslo_concurrency.lockutils [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 33.755s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.885157] env[63202]: DEBUG nova.network.neutron [req-1eb02d81-2104-42e1-b618-55b8116b2baf req-d76b9ad4-fe47-445d-99b8-21283da72168 service nova] [instance: a97cff65-e350-480e-9891-3317fe05be47] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 725.916554] env[63202]: DEBUG nova.network.neutron [-] [instance: a97cff65-e350-480e-9891-3317fe05be47] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.955692] env[63202]: DEBUG nova.network.neutron [req-1eb02d81-2104-42e1-b618-55b8116b2baf req-d76b9ad4-fe47-445d-99b8-21283da72168 service nova] [instance: a97cff65-e350-480e-9891-3317fe05be47] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.162515] env[63202]: DEBUG nova.compute.utils [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 726.164407] env[63202]: DEBUG nova.compute.manager [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 726.164407] env[63202]: DEBUG nova.network.neutron [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 726.210795] env[63202]: DEBUG nova.policy [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e8f61e346d654c4399b59c59e94ed227', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '777b9af00b5145a99c726dcc9b5f8c22', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 726.419803] env[63202]: INFO nova.compute.manager [-] [instance: a97cff65-e350-480e-9891-3317fe05be47] Took 1.02 seconds to deallocate network for instance. [ 726.422095] env[63202]: DEBUG nova.compute.claims [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] [instance: a97cff65-e350-480e-9891-3317fe05be47] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 726.422239] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.458800] env[63202]: DEBUG oslo_concurrency.lockutils [req-1eb02d81-2104-42e1-b618-55b8116b2baf req-d76b9ad4-fe47-445d-99b8-21283da72168 service nova] Releasing lock "refresh_cache-a97cff65-e350-480e-9891-3317fe05be47" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.459252] env[63202]: DEBUG nova.compute.manager [req-1eb02d81-2104-42e1-b618-55b8116b2baf req-d76b9ad4-fe47-445d-99b8-21283da72168 service nova] [instance: a97cff65-e350-480e-9891-3317fe05be47] Received event network-vif-deleted-d12cd099-0585-4a87-9039-680bda744ef9 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 726.502207] env[63202]: DEBUG nova.network.neutron [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] Successfully created port: b68041ef-6834-4ab9-b951-ad2d1b91e460 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 726.543922] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39e511b3-c65e-44fb-8bac-a8bc789ecf23 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.551924] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6459a237-991e-46eb-ac68-4e2fba7515d0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.581194] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abef16e9-b8de-40d4-8680-037b08754087 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.588726] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d4d5f68-4b77-4d00-8e44-9a843918d181 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.601562] env[63202]: DEBUG nova.compute.provider_tree [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 726.667372] env[63202]: DEBUG nova.compute.manager [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 726.764253] env[63202]: DEBUG nova.network.neutron [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] Successfully created port: e0584e3c-e23d-48ab-8ccd-b74fc7ceb17f {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 727.105149] env[63202]: DEBUG nova.scheduler.client.report [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 727.570092] env[63202]: DEBUG nova.compute.manager [req-5b712df7-181c-48bc-b445-a34a37c34eba req-5de04a77-b3d7-47a6-8fd9-c40cf405a515 service nova] [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] Received event network-changed-b68041ef-6834-4ab9-b951-ad2d1b91e460 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 727.570298] env[63202]: DEBUG nova.compute.manager [req-5b712df7-181c-48bc-b445-a34a37c34eba req-5de04a77-b3d7-47a6-8fd9-c40cf405a515 service nova] [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] Refreshing instance network info cache due to event network-changed-b68041ef-6834-4ab9-b951-ad2d1b91e460. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 727.570506] env[63202]: DEBUG oslo_concurrency.lockutils [req-5b712df7-181c-48bc-b445-a34a37c34eba req-5de04a77-b3d7-47a6-8fd9-c40cf405a515 service nova] Acquiring lock "refresh_cache-4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 727.570659] env[63202]: DEBUG oslo_concurrency.lockutils [req-5b712df7-181c-48bc-b445-a34a37c34eba req-5de04a77-b3d7-47a6-8fd9-c40cf405a515 service nova] Acquired lock "refresh_cache-4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.570788] env[63202]: DEBUG nova.network.neutron [req-5b712df7-181c-48bc-b445-a34a37c34eba req-5de04a77-b3d7-47a6-8fd9-c40cf405a515 service nova] [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] Refreshing network info cache for port b68041ef-6834-4ab9-b951-ad2d1b91e460 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 727.610723] env[63202]: DEBUG oslo_concurrency.lockutils [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.956s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 727.611418] env[63202]: ERROR nova.compute.manager [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5d0c4734-7baa-410e-9f3e-91d875f0248b, please check neutron logs for more information. [ 727.611418] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] Traceback (most recent call last): [ 727.611418] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 727.611418] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] self.driver.spawn(context, instance, image_meta, [ 727.611418] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 727.611418] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] self._vmops.spawn(context, instance, image_meta, injected_files, [ 727.611418] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 727.611418] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] vm_ref = self.build_virtual_machine(instance, [ 727.611418] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 727.611418] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] vif_infos = vmwarevif.get_vif_info(self._session, [ 727.611418] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 727.611865] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] for vif in network_info: [ 727.611865] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 727.611865] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] return self._sync_wrapper(fn, *args, **kwargs) [ 727.611865] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 727.611865] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] self.wait() [ 727.611865] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 727.611865] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] self[:] = self._gt.wait() [ 727.611865] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 727.611865] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] return self._exit_event.wait() [ 727.611865] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 727.611865] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] result = hub.switch() [ 727.611865] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 727.611865] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] return self.greenlet.switch() [ 727.612356] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 727.612356] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] result = function(*args, **kwargs) [ 727.612356] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 727.612356] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] return func(*args, **kwargs) [ 727.612356] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 727.612356] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] raise e [ 727.612356] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 727.612356] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] nwinfo = self.network_api.allocate_for_instance( [ 727.612356] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 727.612356] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] created_port_ids = self._update_ports_for_instance( [ 727.612356] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 727.612356] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] with excutils.save_and_reraise_exception(): [ 727.612356] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 727.612735] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] self.force_reraise() [ 727.612735] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 727.612735] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] raise self.value [ 727.612735] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 727.612735] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] updated_port = self._update_port( [ 727.612735] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 727.612735] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] _ensure_no_port_binding_failure(port) [ 727.612735] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 727.612735] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] raise exception.PortBindingFailed(port_id=port['id']) [ 727.612735] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] nova.exception.PortBindingFailed: Binding failed for port 5d0c4734-7baa-410e-9f3e-91d875f0248b, please check neutron logs for more information. [ 727.612735] env[63202]: ERROR nova.compute.manager [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] [ 727.613069] env[63202]: DEBUG nova.compute.utils [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] Binding failed for port 5d0c4734-7baa-410e-9f3e-91d875f0248b, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 727.615118] env[63202]: DEBUG nova.compute.manager [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] Build of instance ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88 was re-scheduled: Binding failed for port 5d0c4734-7baa-410e-9f3e-91d875f0248b, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 727.615534] env[63202]: DEBUG nova.compute.manager [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 727.615754] env[63202]: DEBUG oslo_concurrency.lockutils [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Acquiring lock "refresh_cache-ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 727.615893] env[63202]: DEBUG oslo_concurrency.lockutils [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Acquired lock "refresh_cache-ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.616060] env[63202]: DEBUG nova.network.neutron [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 727.619855] env[63202]: DEBUG oslo_concurrency.lockutils [None req-9c0370f1-864c-44a4-bb17-94eac1e1a874 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 35.541s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 727.619855] env[63202]: DEBUG nova.objects.instance [None req-9c0370f1-864c-44a4-bb17-94eac1e1a874 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Lazy-loading 'resources' on Instance uuid e9b604fb-1d88-425c-92b7-008ddaec7c1a {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 727.676817] env[63202]: DEBUG nova.compute.manager [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 727.726973] env[63202]: DEBUG nova.virt.hardware [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 727.726973] env[63202]: DEBUG nova.virt.hardware [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 727.726973] env[63202]: DEBUG nova.virt.hardware [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 727.726973] env[63202]: DEBUG nova.virt.hardware [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 727.727219] env[63202]: DEBUG nova.virt.hardware [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 727.727253] env[63202]: DEBUG nova.virt.hardware [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 727.727460] env[63202]: DEBUG nova.virt.hardware [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 727.727617] env[63202]: DEBUG nova.virt.hardware [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 727.727778] env[63202]: DEBUG nova.virt.hardware [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 727.727937] env[63202]: DEBUG nova.virt.hardware [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 727.728123] env[63202]: DEBUG nova.virt.hardware [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 727.729047] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-073e87a6-d55d-4a1b-85e8-9f467137374a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.737599] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2029b22d-f3cc-42df-85ae-86b4e7ac6f26 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.774014] env[63202]: ERROR nova.compute.manager [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b68041ef-6834-4ab9-b951-ad2d1b91e460, please check neutron logs for more information. [ 727.774014] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 727.774014] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 727.774014] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 727.774014] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 727.774014] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 727.774014] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 727.774014] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 727.774014] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 727.774014] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 727.774014] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 727.774014] env[63202]: ERROR nova.compute.manager raise self.value [ 727.774014] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 727.774014] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 727.774014] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 727.774014] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 727.774457] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 727.774457] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 727.774457] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b68041ef-6834-4ab9-b951-ad2d1b91e460, please check neutron logs for more information. [ 727.774457] env[63202]: ERROR nova.compute.manager [ 727.774457] env[63202]: Traceback (most recent call last): [ 727.774457] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 727.774457] env[63202]: listener.cb(fileno) [ 727.774457] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 727.774457] env[63202]: result = function(*args, **kwargs) [ 727.774457] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 727.774457] env[63202]: return func(*args, **kwargs) [ 727.774457] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 727.774457] env[63202]: raise e [ 727.774457] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 727.774457] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 727.774457] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 727.774457] env[63202]: created_port_ids = self._update_ports_for_instance( [ 727.774457] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 727.774457] env[63202]: with excutils.save_and_reraise_exception(): [ 727.774457] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 727.774457] env[63202]: self.force_reraise() [ 727.774457] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 727.774457] env[63202]: raise self.value [ 727.774457] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 727.774457] env[63202]: updated_port = self._update_port( [ 727.774457] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 727.774457] env[63202]: _ensure_no_port_binding_failure(port) [ 727.774457] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 727.774457] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 727.775402] env[63202]: nova.exception.PortBindingFailed: Binding failed for port b68041ef-6834-4ab9-b951-ad2d1b91e460, please check neutron logs for more information. [ 727.775402] env[63202]: Removing descriptor: 20 [ 727.775402] env[63202]: ERROR nova.compute.manager [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b68041ef-6834-4ab9-b951-ad2d1b91e460, please check neutron logs for more information. [ 727.775402] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] Traceback (most recent call last): [ 727.775402] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 727.775402] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] yield resources [ 727.775402] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 727.775402] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] self.driver.spawn(context, instance, image_meta, [ 727.775402] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 727.775402] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 727.775402] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 727.775402] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] vm_ref = self.build_virtual_machine(instance, [ 727.775945] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 727.775945] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] vif_infos = vmwarevif.get_vif_info(self._session, [ 727.775945] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 727.775945] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] for vif in network_info: [ 727.775945] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 727.775945] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] return self._sync_wrapper(fn, *args, **kwargs) [ 727.775945] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 727.775945] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] self.wait() [ 727.775945] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 727.775945] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] self[:] = self._gt.wait() [ 727.775945] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 727.775945] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] return self._exit_event.wait() [ 727.775945] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 727.776503] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] result = hub.switch() [ 727.776503] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 727.776503] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] return self.greenlet.switch() [ 727.776503] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 727.776503] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] result = function(*args, **kwargs) [ 727.776503] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 727.776503] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] return func(*args, **kwargs) [ 727.776503] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 727.776503] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] raise e [ 727.776503] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 727.776503] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] nwinfo = self.network_api.allocate_for_instance( [ 727.776503] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 727.776503] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] created_port_ids = self._update_ports_for_instance( [ 727.776825] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 727.776825] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] with excutils.save_and_reraise_exception(): [ 727.776825] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 727.776825] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] self.force_reraise() [ 727.776825] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 727.776825] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] raise self.value [ 727.776825] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 727.776825] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] updated_port = self._update_port( [ 727.776825] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 727.776825] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] _ensure_no_port_binding_failure(port) [ 727.776825] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 727.776825] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] raise exception.PortBindingFailed(port_id=port['id']) [ 727.777318] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] nova.exception.PortBindingFailed: Binding failed for port b68041ef-6834-4ab9-b951-ad2d1b91e460, please check neutron logs for more information. [ 727.777318] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] [ 727.777318] env[63202]: INFO nova.compute.manager [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] Terminating instance [ 727.778189] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Acquiring lock "refresh_cache-4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.105438] env[63202]: DEBUG nova.network.neutron [req-5b712df7-181c-48bc-b445-a34a37c34eba req-5de04a77-b3d7-47a6-8fd9-c40cf405a515 service nova] [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 728.136683] env[63202]: DEBUG nova.network.neutron [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 728.198116] env[63202]: DEBUG nova.network.neutron [req-5b712df7-181c-48bc-b445-a34a37c34eba req-5de04a77-b3d7-47a6-8fd9-c40cf405a515 service nova] [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.232196] env[63202]: DEBUG nova.network.neutron [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.468928] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28e109f3-88c4-4b8e-bbcf-22f1d1a6f08f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.476085] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb308687-99a2-4a90-86ce-fb1ad49b11c5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.506448] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce3253ac-803e-46a3-8649-7031291c838c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.513678] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7970842f-46d7-4070-872a-d3f5ecf8296e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.526426] env[63202]: DEBUG nova.compute.provider_tree [None req-9c0370f1-864c-44a4-bb17-94eac1e1a874 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 728.701010] env[63202]: DEBUG oslo_concurrency.lockutils [req-5b712df7-181c-48bc-b445-a34a37c34eba req-5de04a77-b3d7-47a6-8fd9-c40cf405a515 service nova] Releasing lock "refresh_cache-4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 728.701434] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Acquired lock "refresh_cache-4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.701624] env[63202]: DEBUG nova.network.neutron [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 728.734120] env[63202]: DEBUG oslo_concurrency.lockutils [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Releasing lock "refresh_cache-ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 728.734254] env[63202]: DEBUG nova.compute.manager [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 728.734432] env[63202]: DEBUG nova.compute.manager [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 728.734590] env[63202]: DEBUG nova.network.neutron [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 728.749744] env[63202]: DEBUG nova.network.neutron [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 729.029531] env[63202]: DEBUG nova.scheduler.client.report [None req-9c0370f1-864c-44a4-bb17-94eac1e1a874 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 729.236507] env[63202]: DEBUG nova.network.neutron [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 729.251802] env[63202]: DEBUG nova.network.neutron [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.320917] env[63202]: DEBUG nova.network.neutron [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.534873] env[63202]: DEBUG oslo_concurrency.lockutils [None req-9c0370f1-864c-44a4-bb17-94eac1e1a874 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.917s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.537781] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 36.102s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 729.538983] env[63202]: DEBUG nova.objects.instance [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63202) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 729.557025] env[63202]: INFO nova.scheduler.client.report [None req-9c0370f1-864c-44a4-bb17-94eac1e1a874 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Deleted allocations for instance e9b604fb-1d88-425c-92b7-008ddaec7c1a [ 729.593599] env[63202]: DEBUG nova.compute.manager [req-468ce947-3392-4d0a-b42d-47124c933ac1 req-26a366b1-867a-4ca5-9cb1-f47748aa9f6b service nova] [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] Received event network-vif-deleted-b68041ef-6834-4ab9-b951-ad2d1b91e460 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 729.754630] env[63202]: INFO nova.compute.manager [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88] Took 1.02 seconds to deallocate network for instance. [ 729.822879] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Releasing lock "refresh_cache-4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.823314] env[63202]: DEBUG nova.compute.manager [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 729.823502] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 729.823781] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ea07bb12-8a8d-4080-9388-3b2203059b24 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.832791] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6160f27c-4f58-4ed7-9969-4f559d9530d2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.856750] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1 could not be found. [ 729.856818] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 729.856991] env[63202]: INFO nova.compute.manager [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] Took 0.03 seconds to destroy the instance on the hypervisor. [ 729.857249] env[63202]: DEBUG oslo.service.loopingcall [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 729.857481] env[63202]: DEBUG nova.compute.manager [-] [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 729.857574] env[63202]: DEBUG nova.network.neutron [-] [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 729.905117] env[63202]: DEBUG nova.network.neutron [-] [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 730.064795] env[63202]: DEBUG oslo_concurrency.lockutils [None req-9c0370f1-864c-44a4-bb17-94eac1e1a874 tempest-ServerShowV254Test-594392646 tempest-ServerShowV254Test-594392646-project-member] Lock "e9b604fb-1d88-425c-92b7-008ddaec7c1a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 41.687s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 730.547299] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d104275b-10d3-4f12-b101-ecd644e74d59 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.009s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 730.548455] env[63202]: DEBUG oslo_concurrency.lockutils [None req-31c518cb-d896-429a-88e7-bb67df082013 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.530s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.548864] env[63202]: DEBUG nova.objects.instance [None req-31c518cb-d896-429a-88e7-bb67df082013 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Lazy-loading 'resources' on Instance uuid c8d27e19-0f3c-4cc3-826b-e07b3a635e5e {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 730.703889] env[63202]: DEBUG nova.network.neutron [-] [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.786923] env[63202]: INFO nova.scheduler.client.report [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Deleted allocations for instance ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88 [ 731.206465] env[63202]: INFO nova.compute.manager [-] [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] Took 1.35 seconds to deallocate network for instance. [ 731.210852] env[63202]: DEBUG nova.compute.claims [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 731.211044] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 731.295191] env[63202]: DEBUG oslo_concurrency.lockutils [None req-eeeb6948-5417-4768-960c-652a6738c14f tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Lock "ef7a2a36-4e85-40b3-bb6d-9bc5ee563c88" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.880s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 731.383127] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98415d4b-e25e-46ab-9ee1-260c1237219a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.391508] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6c49952-d4b2-48b6-8978-916c42ac3236 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.421946] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-012ca5f5-0e2c-440e-a5b7-fdf6c1cb5320 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.429365] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f0469d1-4bd3-456a-9c6a-b513f2253782 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.442792] env[63202]: DEBUG nova.compute.provider_tree [None req-31c518cb-d896-429a-88e7-bb67df082013 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 731.799577] env[63202]: DEBUG nova.compute.manager [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 731.947463] env[63202]: DEBUG nova.scheduler.client.report [None req-31c518cb-d896-429a-88e7-bb67df082013 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 732.325240] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 732.452459] env[63202]: DEBUG oslo_concurrency.lockutils [None req-31c518cb-d896-429a-88e7-bb67df082013 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.904s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.455085] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 23.216s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 732.473267] env[63202]: INFO nova.scheduler.client.report [None req-31c518cb-d896-429a-88e7-bb67df082013 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Deleted allocations for instance c8d27e19-0f3c-4cc3-826b-e07b3a635e5e [ 732.980987] env[63202]: DEBUG oslo_concurrency.lockutils [None req-31c518cb-d896-429a-88e7-bb67df082013 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Lock "c8d27e19-0f3c-4cc3-826b-e07b3a635e5e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.662s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.186133] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8fdaf019-1619-4801-8841-bd0e7ea87519 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Acquiring lock "8cd0415d-dab9-4554-966c-1ea8d822deff" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.186133] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8fdaf019-1619-4801-8841-bd0e7ea87519 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Lock "8cd0415d-dab9-4554-966c-1ea8d822deff" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.186133] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8fdaf019-1619-4801-8841-bd0e7ea87519 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Acquiring lock "8cd0415d-dab9-4554-966c-1ea8d822deff-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.186133] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8fdaf019-1619-4801-8841-bd0e7ea87519 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Lock "8cd0415d-dab9-4554-966c-1ea8d822deff-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.186312] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8fdaf019-1619-4801-8841-bd0e7ea87519 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Lock "8cd0415d-dab9-4554-966c-1ea8d822deff-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.188788] env[63202]: INFO nova.compute.manager [None req-8fdaf019-1619-4801-8841-bd0e7ea87519 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: 8cd0415d-dab9-4554-966c-1ea8d822deff] Terminating instance [ 733.190945] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8fdaf019-1619-4801-8841-bd0e7ea87519 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Acquiring lock "refresh_cache-8cd0415d-dab9-4554-966c-1ea8d822deff" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 733.190945] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8fdaf019-1619-4801-8841-bd0e7ea87519 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Acquired lock "refresh_cache-8cd0415d-dab9-4554-966c-1ea8d822deff" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 733.191038] env[63202]: DEBUG nova.network.neutron [None req-8fdaf019-1619-4801-8841-bd0e7ea87519 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: 8cd0415d-dab9-4554-966c-1ea8d822deff] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 733.280698] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d28ee6c-addf-4051-b82f-8124fda4b86b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.288329] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb25909a-ff48-4c2b-9dd3-17830d34d477 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.317868] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ea9782e-79f5-4bee-be8e-ced2a2018f0c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.325076] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5842df4-131a-4c75-bcc5-5a084fc80c3e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.338638] env[63202]: DEBUG nova.compute.provider_tree [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 733.488211] env[63202]: DEBUG oslo_concurrency.lockutils [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Acquiring lock "574ddcba-db96-4bd1-bb9e-3f8846c13430" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.488211] env[63202]: DEBUG oslo_concurrency.lockutils [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Lock "574ddcba-db96-4bd1-bb9e-3f8846c13430" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.708653] env[63202]: DEBUG nova.network.neutron [None req-8fdaf019-1619-4801-8841-bd0e7ea87519 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: 8cd0415d-dab9-4554-966c-1ea8d822deff] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 733.767501] env[63202]: DEBUG nova.network.neutron [None req-8fdaf019-1619-4801-8841-bd0e7ea87519 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: 8cd0415d-dab9-4554-966c-1ea8d822deff] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.843052] env[63202]: DEBUG nova.scheduler.client.report [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 734.269941] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8fdaf019-1619-4801-8841-bd0e7ea87519 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Releasing lock "refresh_cache-8cd0415d-dab9-4554-966c-1ea8d822deff" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 734.270384] env[63202]: DEBUG nova.compute.manager [None req-8fdaf019-1619-4801-8841-bd0e7ea87519 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: 8cd0415d-dab9-4554-966c-1ea8d822deff] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 734.270575] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-8fdaf019-1619-4801-8841-bd0e7ea87519 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: 8cd0415d-dab9-4554-966c-1ea8d822deff] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 734.271516] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fe5d6dd-b256-40ab-9d27-086418ece658 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.279482] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fdaf019-1619-4801-8841-bd0e7ea87519 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: 8cd0415d-dab9-4554-966c-1ea8d822deff] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 734.279693] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0a5cf534-2f41-4bd5-a722-623dfd040953 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.285574] env[63202]: DEBUG oslo_vmware.api [None req-8fdaf019-1619-4801-8841-bd0e7ea87519 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Waiting for the task: (returnval){ [ 734.285574] env[63202]: value = "task-1385373" [ 734.285574] env[63202]: _type = "Task" [ 734.285574] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.293055] env[63202]: DEBUG oslo_vmware.api [None req-8fdaf019-1619-4801-8841-bd0e7ea87519 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385373, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.348065] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.893s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 734.348741] env[63202]: ERROR nova.compute.manager [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0cb05e8b-98bc-4258-bbfa-da06bc6f66cf, please check neutron logs for more information. [ 734.348741] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] Traceback (most recent call last): [ 734.348741] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 734.348741] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] self.driver.spawn(context, instance, image_meta, [ 734.348741] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 734.348741] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 734.348741] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 734.348741] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] vm_ref = self.build_virtual_machine(instance, [ 734.348741] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 734.348741] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] vif_infos = vmwarevif.get_vif_info(self._session, [ 734.348741] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 734.349217] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] for vif in network_info: [ 734.349217] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 734.349217] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] return self._sync_wrapper(fn, *args, **kwargs) [ 734.349217] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 734.349217] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] self.wait() [ 734.349217] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 734.349217] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] self[:] = self._gt.wait() [ 734.349217] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 734.349217] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] return self._exit_event.wait() [ 734.349217] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 734.349217] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] current.throw(*self._exc) [ 734.349217] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 734.349217] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] result = function(*args, **kwargs) [ 734.349756] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 734.349756] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] return func(*args, **kwargs) [ 734.349756] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 734.349756] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] raise e [ 734.349756] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 734.349756] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] nwinfo = self.network_api.allocate_for_instance( [ 734.349756] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 734.349756] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] created_port_ids = self._update_ports_for_instance( [ 734.349756] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 734.349756] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] with excutils.save_and_reraise_exception(): [ 734.349756] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 734.349756] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] self.force_reraise() [ 734.349756] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 734.350321] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] raise self.value [ 734.350321] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 734.350321] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] updated_port = self._update_port( [ 734.350321] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 734.350321] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] _ensure_no_port_binding_failure(port) [ 734.350321] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 734.350321] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] raise exception.PortBindingFailed(port_id=port['id']) [ 734.350321] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] nova.exception.PortBindingFailed: Binding failed for port 0cb05e8b-98bc-4258-bbfa-da06bc6f66cf, please check neutron logs for more information. [ 734.350321] env[63202]: ERROR nova.compute.manager [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] [ 734.350321] env[63202]: DEBUG nova.compute.utils [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] Binding failed for port 0cb05e8b-98bc-4258-bbfa-da06bc6f66cf, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 734.351192] env[63202]: DEBUG nova.compute.manager [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] Build of instance 864e1f52-f77e-4596-a5b5-dff3679f911f was re-scheduled: Binding failed for port 0cb05e8b-98bc-4258-bbfa-da06bc6f66cf, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 734.351579] env[63202]: DEBUG nova.compute.manager [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 734.351801] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] Acquiring lock "refresh_cache-864e1f52-f77e-4596-a5b5-dff3679f911f" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.351945] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] Acquired lock "refresh_cache-864e1f52-f77e-4596-a5b5-dff3679f911f" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.352117] env[63202]: DEBUG nova.network.neutron [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 734.353073] env[63202]: DEBUG oslo_concurrency.lockutils [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 23.197s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 734.796807] env[63202]: DEBUG oslo_vmware.api [None req-8fdaf019-1619-4801-8841-bd0e7ea87519 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385373, 'name': PowerOffVM_Task, 'duration_secs': 0.112822} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.797072] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fdaf019-1619-4801-8841-bd0e7ea87519 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: 8cd0415d-dab9-4554-966c-1ea8d822deff] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 734.797239] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-8fdaf019-1619-4801-8841-bd0e7ea87519 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: 8cd0415d-dab9-4554-966c-1ea8d822deff] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 734.797463] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-33e09f88-f4ce-4f00-94b2-57fe4ce50071 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.820253] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-8fdaf019-1619-4801-8841-bd0e7ea87519 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: 8cd0415d-dab9-4554-966c-1ea8d822deff] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 734.820451] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-8fdaf019-1619-4801-8841-bd0e7ea87519 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: 8cd0415d-dab9-4554-966c-1ea8d822deff] Deleting contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 734.820625] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fdaf019-1619-4801-8841-bd0e7ea87519 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Deleting the datastore file [datastore1] 8cd0415d-dab9-4554-966c-1ea8d822deff {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 734.820873] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b2d663a8-c5cc-42a6-a480-fe0beda49d43 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.826764] env[63202]: DEBUG oslo_vmware.api [None req-8fdaf019-1619-4801-8841-bd0e7ea87519 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Waiting for the task: (returnval){ [ 734.826764] env[63202]: value = "task-1385375" [ 734.826764] env[63202]: _type = "Task" [ 734.826764] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.834080] env[63202]: DEBUG oslo_vmware.api [None req-8fdaf019-1619-4801-8841-bd0e7ea87519 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385375, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.877077] env[63202]: DEBUG nova.network.neutron [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 735.112970] env[63202]: DEBUG nova.network.neutron [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.200765] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61aa9f33-2ed4-4a06-88f4-debda0828261 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.207952] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc9a85a8-f3b5-4849-b95e-651517821eee {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.237672] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b4aae99-ca33-4dd3-a912-b3f20a351099 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.245252] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15d43f6d-cecc-451f-88a9-f2f3b76ca227 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.261226] env[63202]: DEBUG nova.compute.provider_tree [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 735.336987] env[63202]: DEBUG oslo_vmware.api [None req-8fdaf019-1619-4801-8841-bd0e7ea87519 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Task: {'id': task-1385375, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.09253} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.337798] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fdaf019-1619-4801-8841-bd0e7ea87519 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 735.337798] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-8fdaf019-1619-4801-8841-bd0e7ea87519 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: 8cd0415d-dab9-4554-966c-1ea8d822deff] Deleted contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 735.337798] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-8fdaf019-1619-4801-8841-bd0e7ea87519 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: 8cd0415d-dab9-4554-966c-1ea8d822deff] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 735.337798] env[63202]: INFO nova.compute.manager [None req-8fdaf019-1619-4801-8841-bd0e7ea87519 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] [instance: 8cd0415d-dab9-4554-966c-1ea8d822deff] Took 1.07 seconds to destroy the instance on the hypervisor. [ 735.338019] env[63202]: DEBUG oslo.service.loopingcall [None req-8fdaf019-1619-4801-8841-bd0e7ea87519 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 735.338207] env[63202]: DEBUG nova.compute.manager [-] [instance: 8cd0415d-dab9-4554-966c-1ea8d822deff] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 735.338263] env[63202]: DEBUG nova.network.neutron [-] [instance: 8cd0415d-dab9-4554-966c-1ea8d822deff] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 735.354283] env[63202]: DEBUG nova.network.neutron [-] [instance: 8cd0415d-dab9-4554-966c-1ea8d822deff] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 735.615412] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] Releasing lock "refresh_cache-864e1f52-f77e-4596-a5b5-dff3679f911f" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 735.615669] env[63202]: DEBUG nova.compute.manager [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 735.615834] env[63202]: DEBUG nova.compute.manager [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 735.615997] env[63202]: DEBUG nova.network.neutron [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 735.632454] env[63202]: DEBUG nova.network.neutron [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 735.764789] env[63202]: DEBUG nova.scheduler.client.report [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 735.856876] env[63202]: DEBUG nova.network.neutron [-] [instance: 8cd0415d-dab9-4554-966c-1ea8d822deff] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.134847] env[63202]: DEBUG nova.network.neutron [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.270067] env[63202]: DEBUG oslo_concurrency.lockutils [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.917s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.270497] env[63202]: ERROR nova.compute.manager [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3cb97f6c-facd-4990-8a29-63073f5d22a7, please check neutron logs for more information. [ 736.270497] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] Traceback (most recent call last): [ 736.270497] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 736.270497] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] self.driver.spawn(context, instance, image_meta, [ 736.270497] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 736.270497] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] self._vmops.spawn(context, instance, image_meta, injected_files, [ 736.270497] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 736.270497] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] vm_ref = self.build_virtual_machine(instance, [ 736.270497] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 736.270497] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] vif_infos = vmwarevif.get_vif_info(self._session, [ 736.270497] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 736.270804] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] for vif in network_info: [ 736.270804] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 736.270804] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] return self._sync_wrapper(fn, *args, **kwargs) [ 736.270804] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 736.270804] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] self.wait() [ 736.270804] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 736.270804] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] self[:] = self._gt.wait() [ 736.270804] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 736.270804] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] return self._exit_event.wait() [ 736.270804] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 736.270804] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] current.throw(*self._exc) [ 736.270804] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 736.270804] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] result = function(*args, **kwargs) [ 736.271142] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 736.271142] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] return func(*args, **kwargs) [ 736.271142] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 736.271142] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] raise e [ 736.271142] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 736.271142] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] nwinfo = self.network_api.allocate_for_instance( [ 736.271142] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 736.271142] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] created_port_ids = self._update_ports_for_instance( [ 736.271142] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 736.271142] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] with excutils.save_and_reraise_exception(): [ 736.271142] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 736.271142] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] self.force_reraise() [ 736.271142] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 736.271460] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] raise self.value [ 736.271460] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 736.271460] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] updated_port = self._update_port( [ 736.271460] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 736.271460] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] _ensure_no_port_binding_failure(port) [ 736.271460] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 736.271460] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] raise exception.PortBindingFailed(port_id=port['id']) [ 736.271460] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] nova.exception.PortBindingFailed: Binding failed for port 3cb97f6c-facd-4990-8a29-63073f5d22a7, please check neutron logs for more information. [ 736.271460] env[63202]: ERROR nova.compute.manager [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] [ 736.271460] env[63202]: DEBUG nova.compute.utils [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] Binding failed for port 3cb97f6c-facd-4990-8a29-63073f5d22a7, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 736.272958] env[63202]: DEBUG oslo_concurrency.lockutils [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 22.461s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 736.276160] env[63202]: DEBUG nova.compute.manager [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] Build of instance eff8282a-b3cd-4038-9f48-94cd4c4afc55 was re-scheduled: Binding failed for port 3cb97f6c-facd-4990-8a29-63073f5d22a7, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 736.276605] env[63202]: DEBUG nova.compute.manager [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 736.276861] env[63202]: DEBUG oslo_concurrency.lockutils [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] Acquiring lock "refresh_cache-eff8282a-b3cd-4038-9f48-94cd4c4afc55" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 736.277047] env[63202]: DEBUG oslo_concurrency.lockutils [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] Acquired lock "refresh_cache-eff8282a-b3cd-4038-9f48-94cd4c4afc55" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.277242] env[63202]: DEBUG nova.network.neutron [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 736.360341] env[63202]: INFO nova.compute.manager [-] [instance: 8cd0415d-dab9-4554-966c-1ea8d822deff] Took 1.02 seconds to deallocate network for instance. [ 736.637084] env[63202]: INFO nova.compute.manager [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] [instance: 864e1f52-f77e-4596-a5b5-dff3679f911f] Took 1.02 seconds to deallocate network for instance. [ 736.804182] env[63202]: DEBUG nova.network.neutron [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 736.853640] env[63202]: DEBUG nova.network.neutron [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.866633] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8fdaf019-1619-4801-8841-bd0e7ea87519 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 737.080124] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff86ebbf-9134-4dcb-926a-6ab62fc19761 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.087455] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acd3b78f-b305-4fae-8005-23d1729a82b2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.116740] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3772635b-96a0-4801-a716-10cbbb7019d5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.123290] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7963fcb7-3a65-445c-a0dc-15c3f5847dbe {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.135953] env[63202]: DEBUG nova.compute.provider_tree [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 737.356754] env[63202]: DEBUG oslo_concurrency.lockutils [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] Releasing lock "refresh_cache-eff8282a-b3cd-4038-9f48-94cd4c4afc55" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 737.357040] env[63202]: DEBUG nova.compute.manager [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 737.357040] env[63202]: DEBUG nova.compute.manager [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 737.358137] env[63202]: DEBUG nova.network.neutron [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 737.371052] env[63202]: DEBUG nova.network.neutron [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 737.640386] env[63202]: DEBUG nova.scheduler.client.report [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 737.666522] env[63202]: INFO nova.scheduler.client.report [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] Deleted allocations for instance 864e1f52-f77e-4596-a5b5-dff3679f911f [ 737.874062] env[63202]: DEBUG nova.network.neutron [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.145649] env[63202]: DEBUG oslo_concurrency.lockutils [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.872s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.146280] env[63202]: ERROR nova.compute.manager [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] [instance: de5f9555-3959-41cd-b63a-db128a001631] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e280fa17-a568-4c0b-9e68-74717594a05a, please check neutron logs for more information. [ 738.146280] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] Traceback (most recent call last): [ 738.146280] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 738.146280] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] self.driver.spawn(context, instance, image_meta, [ 738.146280] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 738.146280] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] self._vmops.spawn(context, instance, image_meta, injected_files, [ 738.146280] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 738.146280] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] vm_ref = self.build_virtual_machine(instance, [ 738.146280] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 738.146280] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] vif_infos = vmwarevif.get_vif_info(self._session, [ 738.146280] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 738.146644] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] for vif in network_info: [ 738.146644] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 738.146644] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] return self._sync_wrapper(fn, *args, **kwargs) [ 738.146644] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 738.146644] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] self.wait() [ 738.146644] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 738.146644] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] self[:] = self._gt.wait() [ 738.146644] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 738.146644] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] return self._exit_event.wait() [ 738.146644] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 738.146644] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] current.throw(*self._exc) [ 738.146644] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 738.146644] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] result = function(*args, **kwargs) [ 738.147025] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 738.147025] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] return func(*args, **kwargs) [ 738.147025] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 738.147025] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] raise e [ 738.147025] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 738.147025] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] nwinfo = self.network_api.allocate_for_instance( [ 738.147025] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 738.147025] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] created_port_ids = self._update_ports_for_instance( [ 738.147025] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 738.147025] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] with excutils.save_and_reraise_exception(): [ 738.147025] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 738.147025] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] self.force_reraise() [ 738.147025] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 738.147387] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] raise self.value [ 738.147387] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 738.147387] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] updated_port = self._update_port( [ 738.147387] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 738.147387] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] _ensure_no_port_binding_failure(port) [ 738.147387] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 738.147387] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] raise exception.PortBindingFailed(port_id=port['id']) [ 738.147387] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] nova.exception.PortBindingFailed: Binding failed for port e280fa17-a568-4c0b-9e68-74717594a05a, please check neutron logs for more information. [ 738.147387] env[63202]: ERROR nova.compute.manager [instance: de5f9555-3959-41cd-b63a-db128a001631] [ 738.147387] env[63202]: DEBUG nova.compute.utils [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] [instance: de5f9555-3959-41cd-b63a-db128a001631] Binding failed for port e280fa17-a568-4c0b-9e68-74717594a05a, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 738.148387] env[63202]: DEBUG oslo_concurrency.lockutils [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 21.859s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.151148] env[63202]: DEBUG nova.compute.manager [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] [instance: de5f9555-3959-41cd-b63a-db128a001631] Build of instance de5f9555-3959-41cd-b63a-db128a001631 was re-scheduled: Binding failed for port e280fa17-a568-4c0b-9e68-74717594a05a, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 738.151562] env[63202]: DEBUG nova.compute.manager [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] [instance: de5f9555-3959-41cd-b63a-db128a001631] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 738.152031] env[63202]: DEBUG oslo_concurrency.lockutils [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] Acquiring lock "refresh_cache-de5f9555-3959-41cd-b63a-db128a001631" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.152031] env[63202]: DEBUG oslo_concurrency.lockutils [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] Acquired lock "refresh_cache-de5f9555-3959-41cd-b63a-db128a001631" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.152114] env[63202]: DEBUG nova.network.neutron [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] [instance: de5f9555-3959-41cd-b63a-db128a001631] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 738.175466] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6aa3b917-0972-4f21-9b1d-d30e2c300076 tempest-ServerDiagnosticsTest-1847538074 tempest-ServerDiagnosticsTest-1847538074-project-member] Lock "864e1f52-f77e-4596-a5b5-dff3679f911f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 115.575s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.376690] env[63202]: INFO nova.compute.manager [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] [instance: eff8282a-b3cd-4038-9f48-94cd4c4afc55] Took 1.02 seconds to deallocate network for instance. [ 738.675557] env[63202]: DEBUG nova.network.neutron [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] [instance: de5f9555-3959-41cd-b63a-db128a001631] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 738.679659] env[63202]: DEBUG nova.compute.manager [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 738.794064] env[63202]: DEBUG nova.network.neutron [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] [instance: de5f9555-3959-41cd-b63a-db128a001631] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.995690] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9850e51a-4030-4aca-9a22-6e9b14f0ae54 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.004954] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a3f54df-4a24-4286-93e9-aba329701524 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.036842] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4d80d48-573e-4bd5-92ac-168c5c862a55 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.044519] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a4111d7-23a3-47cc-8b96-f736161e3238 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.059848] env[63202]: DEBUG nova.compute.provider_tree [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 739.199098] env[63202]: DEBUG oslo_concurrency.lockutils [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 739.297532] env[63202]: DEBUG oslo_concurrency.lockutils [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] Releasing lock "refresh_cache-de5f9555-3959-41cd-b63a-db128a001631" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.297913] env[63202]: DEBUG nova.compute.manager [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 739.298137] env[63202]: DEBUG nova.compute.manager [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] [instance: de5f9555-3959-41cd-b63a-db128a001631] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 739.298308] env[63202]: DEBUG nova.network.neutron [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] [instance: de5f9555-3959-41cd-b63a-db128a001631] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 739.319587] env[63202]: DEBUG nova.network.neutron [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] [instance: de5f9555-3959-41cd-b63a-db128a001631] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 739.411816] env[63202]: INFO nova.scheduler.client.report [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] Deleted allocations for instance eff8282a-b3cd-4038-9f48-94cd4c4afc55 [ 739.565380] env[63202]: DEBUG nova.scheduler.client.report [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 739.825464] env[63202]: DEBUG nova.network.neutron [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] [instance: de5f9555-3959-41cd-b63a-db128a001631] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.922976] env[63202]: DEBUG oslo_concurrency.lockutils [None req-587c09c9-5c4b-48d1-b1d5-3ed21afeab7e tempest-ServersAdminNegativeTestJSON-1602667767 tempest-ServersAdminNegativeTestJSON-1602667767-project-member] Lock "eff8282a-b3cd-4038-9f48-94cd4c4afc55" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 116.869s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.072358] env[63202]: DEBUG oslo_concurrency.lockutils [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.924s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.072967] env[63202]: ERROR nova.compute.manager [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] [instance: 2f87939c-d615-4d13-8695-971e83ca7843] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 232b0dca-d02a-4796-921b-b5dca87b0a1b, please check neutron logs for more information. [ 740.072967] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] Traceback (most recent call last): [ 740.072967] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 740.072967] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] self.driver.spawn(context, instance, image_meta, [ 740.072967] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 740.072967] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] self._vmops.spawn(context, instance, image_meta, injected_files, [ 740.072967] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 740.072967] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] vm_ref = self.build_virtual_machine(instance, [ 740.072967] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 740.072967] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] vif_infos = vmwarevif.get_vif_info(self._session, [ 740.072967] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 740.073315] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] for vif in network_info: [ 740.073315] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 740.073315] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] return self._sync_wrapper(fn, *args, **kwargs) [ 740.073315] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 740.073315] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] self.wait() [ 740.073315] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 740.073315] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] self[:] = self._gt.wait() [ 740.073315] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 740.073315] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] return self._exit_event.wait() [ 740.073315] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 740.073315] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] current.throw(*self._exc) [ 740.073315] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 740.073315] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] result = function(*args, **kwargs) [ 740.073732] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 740.073732] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] return func(*args, **kwargs) [ 740.073732] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 740.073732] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] raise e [ 740.073732] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 740.073732] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] nwinfo = self.network_api.allocate_for_instance( [ 740.073732] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 740.073732] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] created_port_ids = self._update_ports_for_instance( [ 740.073732] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 740.073732] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] with excutils.save_and_reraise_exception(): [ 740.073732] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 740.073732] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] self.force_reraise() [ 740.073732] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 740.074109] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] raise self.value [ 740.074109] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 740.074109] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] updated_port = self._update_port( [ 740.074109] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 740.074109] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] _ensure_no_port_binding_failure(port) [ 740.074109] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 740.074109] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] raise exception.PortBindingFailed(port_id=port['id']) [ 740.074109] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] nova.exception.PortBindingFailed: Binding failed for port 232b0dca-d02a-4796-921b-b5dca87b0a1b, please check neutron logs for more information. [ 740.074109] env[63202]: ERROR nova.compute.manager [instance: 2f87939c-d615-4d13-8695-971e83ca7843] [ 740.074109] env[63202]: DEBUG nova.compute.utils [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] [instance: 2f87939c-d615-4d13-8695-971e83ca7843] Binding failed for port 232b0dca-d02a-4796-921b-b5dca87b0a1b, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 740.074893] env[63202]: DEBUG oslo_concurrency.lockutils [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.551s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.079988] env[63202]: DEBUG nova.compute.manager [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] [instance: 2f87939c-d615-4d13-8695-971e83ca7843] Build of instance 2f87939c-d615-4d13-8695-971e83ca7843 was re-scheduled: Binding failed for port 232b0dca-d02a-4796-921b-b5dca87b0a1b, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 740.080672] env[63202]: DEBUG nova.compute.manager [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] [instance: 2f87939c-d615-4d13-8695-971e83ca7843] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 740.081076] env[63202]: DEBUG oslo_concurrency.lockutils [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] Acquiring lock "refresh_cache-2f87939c-d615-4d13-8695-971e83ca7843" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.081313] env[63202]: DEBUG oslo_concurrency.lockutils [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] Acquired lock "refresh_cache-2f87939c-d615-4d13-8695-971e83ca7843" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.081515] env[63202]: DEBUG nova.network.neutron [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] [instance: 2f87939c-d615-4d13-8695-971e83ca7843] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 740.330325] env[63202]: INFO nova.compute.manager [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] [instance: de5f9555-3959-41cd-b63a-db128a001631] Took 1.03 seconds to deallocate network for instance. [ 740.424468] env[63202]: DEBUG nova.compute.manager [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 740.624950] env[63202]: DEBUG nova.network.neutron [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] [instance: 2f87939c-d615-4d13-8695-971e83ca7843] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 740.818795] env[63202]: DEBUG nova.network.neutron [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] [instance: 2f87939c-d615-4d13-8695-971e83ca7843] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.950023] env[63202]: DEBUG oslo_concurrency.lockutils [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.988797] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40cde1bc-b384-4669-9cae-2a9390f26ee8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.997744] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3090a7e-fa57-4661-85c3-5dddce88168a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.027323] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0a1deaa-d9d5-448f-9a00-8aa90a9ac154 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.034567] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d148906-daf9-4c0d-997e-77ef8b5066f8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.048152] env[63202]: DEBUG nova.compute.provider_tree [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 741.322516] env[63202]: DEBUG oslo_concurrency.lockutils [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] Releasing lock "refresh_cache-2f87939c-d615-4d13-8695-971e83ca7843" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 741.322656] env[63202]: DEBUG nova.compute.manager [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 741.322779] env[63202]: DEBUG nova.compute.manager [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] [instance: 2f87939c-d615-4d13-8695-971e83ca7843] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 741.322943] env[63202]: DEBUG nova.network.neutron [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] [instance: 2f87939c-d615-4d13-8695-971e83ca7843] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 741.348957] env[63202]: DEBUG nova.network.neutron [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] [instance: 2f87939c-d615-4d13-8695-971e83ca7843] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 741.376216] env[63202]: INFO nova.scheduler.client.report [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] Deleted allocations for instance de5f9555-3959-41cd-b63a-db128a001631 [ 741.553536] env[63202]: DEBUG nova.scheduler.client.report [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 741.850529] env[63202]: DEBUG nova.network.neutron [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] [instance: 2f87939c-d615-4d13-8695-971e83ca7843] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.885075] env[63202]: DEBUG oslo_concurrency.lockutils [None req-588c48db-096c-4f6b-afae-899fa6335808 tempest-ServersV294TestFqdnHostnames-519274824 tempest-ServersV294TestFqdnHostnames-519274824-project-member] Lock "de5f9555-3959-41cd-b63a-db128a001631" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.278s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.056923] env[63202]: DEBUG oslo_concurrency.lockutils [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.982s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.057550] env[63202]: ERROR nova.compute.manager [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 501a66a6-4cf1-4ec6-8953-5eb6cbed571e, please check neutron logs for more information. [ 742.057550] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] Traceback (most recent call last): [ 742.057550] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 742.057550] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] self.driver.spawn(context, instance, image_meta, [ 742.057550] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 742.057550] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 742.057550] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 742.057550] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] vm_ref = self.build_virtual_machine(instance, [ 742.057550] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 742.057550] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] vif_infos = vmwarevif.get_vif_info(self._session, [ 742.057550] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 742.057933] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] for vif in network_info: [ 742.057933] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 742.057933] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] return self._sync_wrapper(fn, *args, **kwargs) [ 742.057933] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 742.057933] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] self.wait() [ 742.057933] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 742.057933] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] self[:] = self._gt.wait() [ 742.057933] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 742.057933] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] return self._exit_event.wait() [ 742.057933] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 742.057933] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] current.throw(*self._exc) [ 742.057933] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 742.057933] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] result = function(*args, **kwargs) [ 742.058316] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 742.058316] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] return func(*args, **kwargs) [ 742.058316] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 742.058316] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] raise e [ 742.058316] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 742.058316] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] nwinfo = self.network_api.allocate_for_instance( [ 742.058316] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 742.058316] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] created_port_ids = self._update_ports_for_instance( [ 742.058316] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 742.058316] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] with excutils.save_and_reraise_exception(): [ 742.058316] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 742.058316] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] self.force_reraise() [ 742.058316] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 742.058706] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] raise self.value [ 742.058706] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 742.058706] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] updated_port = self._update_port( [ 742.058706] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 742.058706] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] _ensure_no_port_binding_failure(port) [ 742.058706] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 742.058706] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] raise exception.PortBindingFailed(port_id=port['id']) [ 742.058706] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] nova.exception.PortBindingFailed: Binding failed for port 501a66a6-4cf1-4ec6-8953-5eb6cbed571e, please check neutron logs for more information. [ 742.058706] env[63202]: ERROR nova.compute.manager [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] [ 742.058706] env[63202]: DEBUG nova.compute.utils [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] Binding failed for port 501a66a6-4cf1-4ec6-8953-5eb6cbed571e, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 742.059672] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.672s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 742.064035] env[63202]: DEBUG nova.compute.manager [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] Build of instance 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e was re-scheduled: Binding failed for port 501a66a6-4cf1-4ec6-8953-5eb6cbed571e, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 742.064035] env[63202]: DEBUG nova.compute.manager [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 742.064294] env[63202]: DEBUG oslo_concurrency.lockutils [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Acquiring lock "refresh_cache-8498aa14-fc1a-42e4-a5e6-e22b239a8f7e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.064434] env[63202]: DEBUG oslo_concurrency.lockutils [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Acquired lock "refresh_cache-8498aa14-fc1a-42e4-a5e6-e22b239a8f7e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.064659] env[63202]: DEBUG nova.network.neutron [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 742.355135] env[63202]: INFO nova.compute.manager [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] [instance: 2f87939c-d615-4d13-8695-971e83ca7843] Took 1.03 seconds to deallocate network for instance. [ 742.388602] env[63202]: DEBUG nova.compute.manager [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 742.590097] env[63202]: DEBUG nova.network.neutron [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 742.852703] env[63202]: DEBUG nova.network.neutron [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.918393] env[63202]: DEBUG oslo_concurrency.lockutils [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 743.013953] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6028341c-ccca-4f1f-be99-47c86d36e017 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.022609] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d4e3e6e-5c7c-49bf-94b3-4b7859b1e201 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.058964] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1f7abfb-d68e-490a-9dff-f50b82129e46 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.067306] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-094c195b-1566-4f17-a220-5afff3807f20 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.084126] env[63202]: DEBUG nova.compute.provider_tree [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 743.356450] env[63202]: DEBUG oslo_concurrency.lockutils [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Releasing lock "refresh_cache-8498aa14-fc1a-42e4-a5e6-e22b239a8f7e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.356671] env[63202]: DEBUG nova.compute.manager [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 743.356852] env[63202]: DEBUG nova.compute.manager [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 743.357055] env[63202]: DEBUG nova.network.neutron [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 743.393695] env[63202]: DEBUG nova.network.neutron [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 743.398634] env[63202]: INFO nova.scheduler.client.report [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] Deleted allocations for instance 2f87939c-d615-4d13-8695-971e83ca7843 [ 743.587136] env[63202]: DEBUG nova.scheduler.client.report [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 743.897516] env[63202]: DEBUG nova.network.neutron [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.915807] env[63202]: DEBUG oslo_concurrency.lockutils [None req-06ac6a34-6be1-4aff-b46a-5194a405c3c8 tempest-ServerRescueTestJSON-540830590 tempest-ServerRescueTestJSON-540830590-project-member] Lock "2f87939c-d615-4d13-8695-971e83ca7843" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 118.667s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.095391] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.034s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.096013] env[63202]: ERROR nova.compute.manager [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 06496aad-b025-48c2-8436-03d43d1ac899] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ce51d939-c2c1-4116-9d39-577687daaf4f, please check neutron logs for more information. [ 744.096013] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] Traceback (most recent call last): [ 744.096013] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 744.096013] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] self.driver.spawn(context, instance, image_meta, [ 744.096013] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 744.096013] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] self._vmops.spawn(context, instance, image_meta, injected_files, [ 744.096013] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 744.096013] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] vm_ref = self.build_virtual_machine(instance, [ 744.096013] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 744.096013] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] vif_infos = vmwarevif.get_vif_info(self._session, [ 744.096013] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 744.096364] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] for vif in network_info: [ 744.096364] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 744.096364] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] return self._sync_wrapper(fn, *args, **kwargs) [ 744.096364] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 744.096364] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] self.wait() [ 744.096364] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 744.096364] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] self[:] = self._gt.wait() [ 744.096364] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 744.096364] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] return self._exit_event.wait() [ 744.096364] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 744.096364] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] current.throw(*self._exc) [ 744.096364] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 744.096364] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] result = function(*args, **kwargs) [ 744.096742] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 744.096742] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] return func(*args, **kwargs) [ 744.096742] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 744.096742] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] raise e [ 744.096742] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 744.096742] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] nwinfo = self.network_api.allocate_for_instance( [ 744.096742] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 744.096742] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] created_port_ids = self._update_ports_for_instance( [ 744.096742] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 744.096742] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] with excutils.save_and_reraise_exception(): [ 744.096742] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 744.096742] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] self.force_reraise() [ 744.096742] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 744.097143] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] raise self.value [ 744.097143] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 744.097143] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] updated_port = self._update_port( [ 744.097143] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 744.097143] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] _ensure_no_port_binding_failure(port) [ 744.097143] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 744.097143] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] raise exception.PortBindingFailed(port_id=port['id']) [ 744.097143] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] nova.exception.PortBindingFailed: Binding failed for port ce51d939-c2c1-4116-9d39-577687daaf4f, please check neutron logs for more information. [ 744.097143] env[63202]: ERROR nova.compute.manager [instance: 06496aad-b025-48c2-8436-03d43d1ac899] [ 744.097143] env[63202]: DEBUG nova.compute.utils [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 06496aad-b025-48c2-8436-03d43d1ac899] Binding failed for port ce51d939-c2c1-4116-9d39-577687daaf4f, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 744.099681] env[63202]: DEBUG nova.compute.manager [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 06496aad-b025-48c2-8436-03d43d1ac899] Build of instance 06496aad-b025-48c2-8436-03d43d1ac899 was re-scheduled: Binding failed for port ce51d939-c2c1-4116-9d39-577687daaf4f, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 744.100406] env[63202]: DEBUG nova.compute.manager [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 06496aad-b025-48c2-8436-03d43d1ac899] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 744.101114] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Acquiring lock "refresh_cache-06496aad-b025-48c2-8436-03d43d1ac899" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 744.101323] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Acquired lock "refresh_cache-06496aad-b025-48c2-8436-03d43d1ac899" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.101972] env[63202]: DEBUG nova.network.neutron [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 06496aad-b025-48c2-8436-03d43d1ac899] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 744.104715] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.172s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.376695] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] Acquiring lock "57f446ee-edf3-40e7-96bf-8fdce62bb19f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.376695] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] Lock "57f446ee-edf3-40e7-96bf-8fdce62bb19f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.406711] env[63202]: INFO nova.compute.manager [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e] Took 1.05 seconds to deallocate network for instance. [ 744.419735] env[63202]: DEBUG nova.compute.manager [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 744.632857] env[63202]: DEBUG nova.network.neutron [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 06496aad-b025-48c2-8436-03d43d1ac899] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 744.833738] env[63202]: DEBUG nova.network.neutron [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 06496aad-b025-48c2-8436-03d43d1ac899] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.943200] env[63202]: DEBUG oslo_concurrency.lockutils [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 745.046925] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a76555b-f2a3-4186-b419-6edf07c97e84 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.057704] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ea51eff-a10e-485e-ad2d-deacaaa19396 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.093777] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be099a95-a874-40d4-8f13-5978a632bd39 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.101935] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e59ced2c-8184-4ce1-96b9-39f50938050e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.115141] env[63202]: DEBUG nova.compute.provider_tree [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 745.337387] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Releasing lock "refresh_cache-06496aad-b025-48c2-8436-03d43d1ac899" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 745.337387] env[63202]: DEBUG nova.compute.manager [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 745.337597] env[63202]: DEBUG nova.compute.manager [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 06496aad-b025-48c2-8436-03d43d1ac899] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 745.337733] env[63202]: DEBUG nova.network.neutron [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 06496aad-b025-48c2-8436-03d43d1ac899] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 745.358084] env[63202]: DEBUG nova.network.neutron [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 06496aad-b025-48c2-8436-03d43d1ac899] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 745.449053] env[63202]: INFO nova.scheduler.client.report [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Deleted allocations for instance 8498aa14-fc1a-42e4-a5e6-e22b239a8f7e [ 745.618792] env[63202]: DEBUG nova.scheduler.client.report [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 745.862155] env[63202]: DEBUG nova.network.neutron [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 06496aad-b025-48c2-8436-03d43d1ac899] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.959221] env[63202]: DEBUG oslo_concurrency.lockutils [None req-207b68bf-0a46-4ec0-b23e-42341fae06cc tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Lock "8498aa14-fc1a-42e4-a5e6-e22b239a8f7e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 118.132s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.999179] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] Acquiring lock "ac48e21a-72ec-4c9d-8262-5d6cb36ec531" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 745.999179] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] Lock "ac48e21a-72ec-4c9d-8262-5d6cb36ec531" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 746.127021] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.021s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 746.127021] env[63202]: ERROR nova.compute.manager [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 09ee3acd-6615-4372-b295-b69bb039fbce, please check neutron logs for more information. [ 746.127021] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] Traceback (most recent call last): [ 746.127021] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 746.127021] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] self.driver.spawn(context, instance, image_meta, [ 746.127021] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 746.127021] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] self._vmops.spawn(context, instance, image_meta, injected_files, [ 746.127021] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 746.127021] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] vm_ref = self.build_virtual_machine(instance, [ 746.127386] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 746.127386] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] vif_infos = vmwarevif.get_vif_info(self._session, [ 746.127386] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 746.127386] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] for vif in network_info: [ 746.127386] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 746.127386] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] return self._sync_wrapper(fn, *args, **kwargs) [ 746.127386] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 746.127386] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] self.wait() [ 746.127386] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 746.127386] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] self[:] = self._gt.wait() [ 746.127386] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 746.127386] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] return self._exit_event.wait() [ 746.127386] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 746.128086] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] current.throw(*self._exc) [ 746.128086] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 746.128086] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] result = function(*args, **kwargs) [ 746.128086] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 746.128086] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] return func(*args, **kwargs) [ 746.128086] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 746.128086] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] raise e [ 746.128086] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 746.128086] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] nwinfo = self.network_api.allocate_for_instance( [ 746.128086] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 746.128086] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] created_port_ids = self._update_ports_for_instance( [ 746.128086] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 746.128086] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] with excutils.save_and_reraise_exception(): [ 746.128433] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 746.128433] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] self.force_reraise() [ 746.128433] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 746.128433] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] raise self.value [ 746.128433] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 746.128433] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] updated_port = self._update_port( [ 746.128433] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 746.128433] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] _ensure_no_port_binding_failure(port) [ 746.128433] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 746.128433] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] raise exception.PortBindingFailed(port_id=port['id']) [ 746.128433] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] nova.exception.PortBindingFailed: Binding failed for port 09ee3acd-6615-4372-b295-b69bb039fbce, please check neutron logs for more information. [ 746.128433] env[63202]: ERROR nova.compute.manager [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] [ 746.128895] env[63202]: DEBUG nova.compute.utils [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] Binding failed for port 09ee3acd-6615-4372-b295-b69bb039fbce, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 746.128895] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.705s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 746.131076] env[63202]: DEBUG nova.compute.manager [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] Build of instance 8885156e-78f1-45ff-87c9-829f3ab89dca was re-scheduled: Binding failed for port 09ee3acd-6615-4372-b295-b69bb039fbce, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 746.131528] env[63202]: DEBUG nova.compute.manager [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 746.131652] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Acquiring lock "refresh_cache-8885156e-78f1-45ff-87c9-829f3ab89dca" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.131874] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Acquired lock "refresh_cache-8885156e-78f1-45ff-87c9-829f3ab89dca" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.131943] env[63202]: DEBUG nova.network.neutron [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 746.366313] env[63202]: INFO nova.compute.manager [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] [instance: 06496aad-b025-48c2-8436-03d43d1ac899] Took 1.03 seconds to deallocate network for instance. [ 746.463452] env[63202]: DEBUG nova.compute.manager [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 746.673517] env[63202]: DEBUG nova.network.neutron [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 746.869578] env[63202]: DEBUG nova.network.neutron [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.964583] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59194083-4722-4462-8651-f09bea856232 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.975559] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ebc8139-ebb6-4c53-83d2-08677b184a0e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.012013] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.012886] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bacc976d-ba2e-4021-827f-3a99b0e02802 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.022680] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9173be02-b69e-44dc-b72e-a3cf260e9c34 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.036545] env[63202]: DEBUG nova.compute.provider_tree [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 747.374945] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Releasing lock "refresh_cache-8885156e-78f1-45ff-87c9-829f3ab89dca" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 747.377014] env[63202]: DEBUG nova.compute.manager [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 747.377014] env[63202]: DEBUG nova.compute.manager [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 747.377014] env[63202]: DEBUG nova.network.neutron [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 747.399281] env[63202]: INFO nova.scheduler.client.report [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Deleted allocations for instance 06496aad-b025-48c2-8436-03d43d1ac899 [ 747.407655] env[63202]: DEBUG nova.network.neutron [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 747.541938] env[63202]: DEBUG nova.scheduler.client.report [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 747.912143] env[63202]: DEBUG nova.network.neutron [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.912729] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a69053c1-c55a-4763-af2a-be2f85f06e7a tempest-ServerRescueNegativeTestJSON-1923478396 tempest-ServerRescueNegativeTestJSON-1923478396-project-member] Lock "06496aad-b025-48c2-8436-03d43d1ac899" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 118.292s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 748.046509] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.919s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 748.048323] env[63202]: ERROR nova.compute.manager [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] [instance: a97cff65-e350-480e-9891-3317fe05be47] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d12cd099-0585-4a87-9039-680bda744ef9, please check neutron logs for more information. [ 748.048323] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] Traceback (most recent call last): [ 748.048323] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 748.048323] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] self.driver.spawn(context, instance, image_meta, [ 748.048323] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 748.048323] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] self._vmops.spawn(context, instance, image_meta, injected_files, [ 748.048323] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 748.048323] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] vm_ref = self.build_virtual_machine(instance, [ 748.048323] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 748.048323] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] vif_infos = vmwarevif.get_vif_info(self._session, [ 748.048323] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 748.048641] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] for vif in network_info: [ 748.048641] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 748.048641] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] return self._sync_wrapper(fn, *args, **kwargs) [ 748.048641] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 748.048641] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] self.wait() [ 748.048641] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 748.048641] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] self[:] = self._gt.wait() [ 748.048641] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 748.048641] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] return self._exit_event.wait() [ 748.048641] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 748.048641] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] current.throw(*self._exc) [ 748.048641] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 748.048641] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] result = function(*args, **kwargs) [ 748.049035] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 748.049035] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] return func(*args, **kwargs) [ 748.049035] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 748.049035] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] raise e [ 748.049035] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 748.049035] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] nwinfo = self.network_api.allocate_for_instance( [ 748.049035] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 748.049035] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] created_port_ids = self._update_ports_for_instance( [ 748.049035] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 748.049035] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] with excutils.save_and_reraise_exception(): [ 748.049035] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 748.049035] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] self.force_reraise() [ 748.049035] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 748.050396] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] raise self.value [ 748.050396] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 748.050396] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] updated_port = self._update_port( [ 748.050396] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 748.050396] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] _ensure_no_port_binding_failure(port) [ 748.050396] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 748.050396] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] raise exception.PortBindingFailed(port_id=port['id']) [ 748.050396] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] nova.exception.PortBindingFailed: Binding failed for port d12cd099-0585-4a87-9039-680bda744ef9, please check neutron logs for more information. [ 748.050396] env[63202]: ERROR nova.compute.manager [instance: a97cff65-e350-480e-9891-3317fe05be47] [ 748.050396] env[63202]: DEBUG nova.compute.utils [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] [instance: a97cff65-e350-480e-9891-3317fe05be47] Binding failed for port d12cd099-0585-4a87-9039-680bda744ef9, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 748.050686] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.839s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 748.056310] env[63202]: DEBUG nova.compute.manager [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] [instance: a97cff65-e350-480e-9891-3317fe05be47] Build of instance a97cff65-e350-480e-9891-3317fe05be47 was re-scheduled: Binding failed for port d12cd099-0585-4a87-9039-680bda744ef9, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 748.056310] env[63202]: DEBUG nova.compute.manager [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] [instance: a97cff65-e350-480e-9891-3317fe05be47] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 748.056310] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] Acquiring lock "refresh_cache-a97cff65-e350-480e-9891-3317fe05be47" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 748.056310] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] Acquired lock "refresh_cache-a97cff65-e350-480e-9891-3317fe05be47" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.056725] env[63202]: DEBUG nova.network.neutron [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] [instance: a97cff65-e350-480e-9891-3317fe05be47] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 748.415996] env[63202]: INFO nova.compute.manager [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 8885156e-78f1-45ff-87c9-829f3ab89dca] Took 1.04 seconds to deallocate network for instance. [ 748.417858] env[63202]: DEBUG nova.compute.manager [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: fea91da2-186a-44b6-8e20-535e3b123890] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 748.605488] env[63202]: DEBUG nova.network.neutron [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] [instance: a97cff65-e350-480e-9891-3317fe05be47] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 748.812019] env[63202]: DEBUG nova.network.neutron [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] [instance: a97cff65-e350-480e-9891-3317fe05be47] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.943029] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64e5a0ba-04c9-415f-a46a-c7a1b43b622a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.951067] env[63202]: DEBUG oslo_concurrency.lockutils [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 748.954904] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22669f0e-0a94-4208-8a7f-6adc73c51e8c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.000803] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae5a58ce-36bd-46a2-9c39-fd2fdb05b099 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.011518] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d1a044e-5dcc-4c8e-84f9-54ee22c5e70d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.029014] env[63202]: DEBUG nova.compute.provider_tree [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 749.318022] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] Releasing lock "refresh_cache-a97cff65-e350-480e-9891-3317fe05be47" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 749.318022] env[63202]: DEBUG nova.compute.manager [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 749.318022] env[63202]: DEBUG nova.compute.manager [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] [instance: a97cff65-e350-480e-9891-3317fe05be47] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 749.318022] env[63202]: DEBUG nova.network.neutron [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] [instance: a97cff65-e350-480e-9891-3317fe05be47] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 749.346682] env[63202]: DEBUG nova.network.neutron [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] [instance: a97cff65-e350-480e-9891-3317fe05be47] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 749.425695] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] Acquiring lock "69c66a35-3789-4c2f-b7cf-5f4136d53b94" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 749.425882] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] Lock "69c66a35-3789-4c2f-b7cf-5f4136d53b94" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 749.470828] env[63202]: INFO nova.scheduler.client.report [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Deleted allocations for instance 8885156e-78f1-45ff-87c9-829f3ab89dca [ 749.533216] env[63202]: DEBUG nova.scheduler.client.report [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 749.849240] env[63202]: DEBUG nova.network.neutron [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] [instance: a97cff65-e350-480e-9891-3317fe05be47] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.983605] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1c69b6aa-5608-41d6-a1c8-af847c55c565 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Lock "8885156e-78f1-45ff-87c9-829f3ab89dca" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 118.562s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.038353] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.988s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.038988] env[63202]: ERROR nova.compute.manager [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b68041ef-6834-4ab9-b951-ad2d1b91e460, please check neutron logs for more information. [ 750.038988] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] Traceback (most recent call last): [ 750.038988] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 750.038988] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] self.driver.spawn(context, instance, image_meta, [ 750.038988] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 750.038988] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 750.038988] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 750.038988] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] vm_ref = self.build_virtual_machine(instance, [ 750.038988] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 750.038988] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] vif_infos = vmwarevif.get_vif_info(self._session, [ 750.038988] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 750.039370] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] for vif in network_info: [ 750.039370] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 750.039370] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] return self._sync_wrapper(fn, *args, **kwargs) [ 750.039370] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 750.039370] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] self.wait() [ 750.039370] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 750.039370] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] self[:] = self._gt.wait() [ 750.039370] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 750.039370] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] return self._exit_event.wait() [ 750.039370] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 750.039370] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] result = hub.switch() [ 750.039370] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 750.039370] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] return self.greenlet.switch() [ 750.039707] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 750.039707] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] result = function(*args, **kwargs) [ 750.039707] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 750.039707] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] return func(*args, **kwargs) [ 750.039707] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 750.039707] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] raise e [ 750.039707] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 750.039707] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] nwinfo = self.network_api.allocate_for_instance( [ 750.039707] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 750.039707] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] created_port_ids = self._update_ports_for_instance( [ 750.039707] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 750.039707] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] with excutils.save_and_reraise_exception(): [ 750.039707] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 750.040099] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] self.force_reraise() [ 750.040099] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 750.040099] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] raise self.value [ 750.040099] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 750.040099] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] updated_port = self._update_port( [ 750.040099] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 750.040099] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] _ensure_no_port_binding_failure(port) [ 750.040099] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 750.040099] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] raise exception.PortBindingFailed(port_id=port['id']) [ 750.040099] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] nova.exception.PortBindingFailed: Binding failed for port b68041ef-6834-4ab9-b951-ad2d1b91e460, please check neutron logs for more information. [ 750.040099] env[63202]: ERROR nova.compute.manager [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] [ 750.040423] env[63202]: DEBUG nova.compute.utils [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] Binding failed for port b68041ef-6834-4ab9-b951-ad2d1b91e460, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 750.042288] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.717s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.043716] env[63202]: INFO nova.compute.claims [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 750.049272] env[63202]: DEBUG nova.compute.manager [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] Build of instance 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1 was re-scheduled: Binding failed for port b68041ef-6834-4ab9-b951-ad2d1b91e460, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 750.049691] env[63202]: DEBUG nova.compute.manager [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 750.049928] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Acquiring lock "refresh_cache-4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.050175] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Acquired lock "refresh_cache-4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.050243] env[63202]: DEBUG nova.network.neutron [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 750.354631] env[63202]: INFO nova.compute.manager [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] [instance: a97cff65-e350-480e-9891-3317fe05be47] Took 1.04 seconds to deallocate network for instance. [ 750.486776] env[63202]: DEBUG nova.compute.manager [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 750.575799] env[63202]: DEBUG nova.network.neutron [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 750.701258] env[63202]: DEBUG nova.network.neutron [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.020330] env[63202]: DEBUG oslo_concurrency.lockutils [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.205779] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Releasing lock "refresh_cache-4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.205972] env[63202]: DEBUG nova.compute.manager [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 751.206160] env[63202]: DEBUG nova.compute.manager [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 751.206324] env[63202]: DEBUG nova.network.neutron [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 751.226734] env[63202]: DEBUG nova.network.neutron [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 751.348949] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c29189ab-1703-49ef-8f4d-1d62904c4cfa {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.356993] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d36d57e6-286f-407c-a003-24452e172001 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.389059] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c91aabf7-bf2e-4ad4-9636-5a841099873b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.397082] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0c0efea-a851-4a39-ac49-2eca795e9861 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.410449] env[63202]: DEBUG nova.compute.provider_tree [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 751.412389] env[63202]: INFO nova.scheduler.client.report [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] Deleted allocations for instance a97cff65-e350-480e-9891-3317fe05be47 [ 751.618669] env[63202]: DEBUG oslo_concurrency.lockutils [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Acquiring lock "5d9f1803-e3f8-42a5-b6b7-2642463a079e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.618912] env[63202]: DEBUG oslo_concurrency.lockutils [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Lock "5d9f1803-e3f8-42a5-b6b7-2642463a079e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.729010] env[63202]: DEBUG nova.network.neutron [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.920050] env[63202]: DEBUG nova.scheduler.client.report [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 751.923196] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89b6ec04-a52a-4ad0-b05b-3c9b483f7d27 tempest-InstanceActionsV221TestJSON-1084937500 tempest-InstanceActionsV221TestJSON-1084937500-project-member] Lock "a97cff65-e350-480e-9891-3317fe05be47" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 118.426s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.232484] env[63202]: INFO nova.compute.manager [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] [instance: 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1] Took 1.03 seconds to deallocate network for instance. [ 752.425353] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.383s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.425864] env[63202]: DEBUG nova.compute.manager [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 752.428707] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8fdaf019-1619-4801-8841-bd0e7ea87519 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.562s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.430633] env[63202]: DEBUG nova.objects.instance [None req-8fdaf019-1619-4801-8841-bd0e7ea87519 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Lazy-loading 'resources' on Instance uuid 8cd0415d-dab9-4554-966c-1ea8d822deff {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 752.430633] env[63202]: DEBUG nova.compute.manager [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 752.936764] env[63202]: DEBUG nova.compute.utils [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 752.942897] env[63202]: DEBUG nova.compute.manager [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 752.943204] env[63202]: DEBUG nova.network.neutron [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 752.962231] env[63202]: DEBUG oslo_concurrency.lockutils [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.993244] env[63202]: DEBUG nova.policy [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bc538e1e5d494591ae07f83cf5ff700c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1d41989252e14a78b0e562fc48076b27', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 753.272814] env[63202]: INFO nova.scheduler.client.report [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Deleted allocations for instance 4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1 [ 753.298736] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c10de5b-3840-4dcd-879a-769c1d0b7a06 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.312158] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d17e497-c4da-40fb-9f72-ce8b96a119d2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.346779] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e16333d-a8e2-490b-8898-924a559362f1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.354557] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09c42fe6-7259-4a80-8267-178c4635a8f2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.367960] env[63202]: DEBUG nova.compute.provider_tree [None req-8fdaf019-1619-4801-8841-bd0e7ea87519 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 753.447113] env[63202]: DEBUG nova.compute.manager [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 753.585098] env[63202]: DEBUG nova.network.neutron [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] Successfully created port: 0cdc9fcc-a493-46de-b498-4389665eaf90 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 753.786851] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8d7f5949-f7b5-4813-a06e-0403ba33eea6 tempest-ServersTestMultiNic-1302030917 tempest-ServersTestMultiNic-1302030917-project-member] Lock "4d0586a5-60b3-4144-87bf-2fc5f9e9c0b1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.120s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.871160] env[63202]: DEBUG nova.scheduler.client.report [None req-8fdaf019-1619-4801-8841-bd0e7ea87519 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 754.291016] env[63202]: DEBUG nova.compute.manager [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 754.376234] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8fdaf019-1619-4801-8841-bd0e7ea87519 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.947s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 754.378566] env[63202]: DEBUG oslo_concurrency.lockutils [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.180s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 754.380411] env[63202]: INFO nova.compute.claims [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 754.403230] env[63202]: INFO nova.scheduler.client.report [None req-8fdaf019-1619-4801-8841-bd0e7ea87519 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Deleted allocations for instance 8cd0415d-dab9-4554-966c-1ea8d822deff [ 754.459656] env[63202]: DEBUG nova.compute.manager [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 754.492038] env[63202]: DEBUG nova.virt.hardware [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 754.492291] env[63202]: DEBUG nova.virt.hardware [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 754.492447] env[63202]: DEBUG nova.virt.hardware [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 754.492662] env[63202]: DEBUG nova.virt.hardware [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 754.492823] env[63202]: DEBUG nova.virt.hardware [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 754.492966] env[63202]: DEBUG nova.virt.hardware [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 754.493221] env[63202]: DEBUG nova.virt.hardware [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 754.493377] env[63202]: DEBUG nova.virt.hardware [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 754.493540] env[63202]: DEBUG nova.virt.hardware [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 754.493699] env[63202]: DEBUG nova.virt.hardware [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 754.493867] env[63202]: DEBUG nova.virt.hardware [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 754.494796] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b843fcc0-7aab-44a3-84f4-0d5cc895bf55 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.503148] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c27641f-555a-49e6-bbce-912ed9c8108e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.616903] env[63202]: DEBUG nova.compute.manager [req-752c2780-1e28-4b15-8d48-3889faa5795e req-cb38e587-7be5-409e-84cf-fa71661fe530 service nova] [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] Received event network-changed-0cdc9fcc-a493-46de-b498-4389665eaf90 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 754.617261] env[63202]: DEBUG nova.compute.manager [req-752c2780-1e28-4b15-8d48-3889faa5795e req-cb38e587-7be5-409e-84cf-fa71661fe530 service nova] [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] Refreshing instance network info cache due to event network-changed-0cdc9fcc-a493-46de-b498-4389665eaf90. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 754.617353] env[63202]: DEBUG oslo_concurrency.lockutils [req-752c2780-1e28-4b15-8d48-3889faa5795e req-cb38e587-7be5-409e-84cf-fa71661fe530 service nova] Acquiring lock "refresh_cache-611fead3-3d1b-41e4-9579-7ad1a1b754c3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 754.617496] env[63202]: DEBUG oslo_concurrency.lockutils [req-752c2780-1e28-4b15-8d48-3889faa5795e req-cb38e587-7be5-409e-84cf-fa71661fe530 service nova] Acquired lock "refresh_cache-611fead3-3d1b-41e4-9579-7ad1a1b754c3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.617653] env[63202]: DEBUG nova.network.neutron [req-752c2780-1e28-4b15-8d48-3889faa5795e req-cb38e587-7be5-409e-84cf-fa71661fe530 service nova] [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] Refreshing network info cache for port 0cdc9fcc-a493-46de-b498-4389665eaf90 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 754.812378] env[63202]: DEBUG oslo_concurrency.lockutils [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 754.864996] env[63202]: ERROR nova.compute.manager [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0cdc9fcc-a493-46de-b498-4389665eaf90, please check neutron logs for more information. [ 754.864996] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 754.864996] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 754.864996] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 754.864996] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 754.864996] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 754.864996] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 754.864996] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 754.864996] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 754.864996] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 754.864996] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 754.864996] env[63202]: ERROR nova.compute.manager raise self.value [ 754.864996] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 754.864996] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 754.864996] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 754.864996] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 754.865657] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 754.865657] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 754.865657] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0cdc9fcc-a493-46de-b498-4389665eaf90, please check neutron logs for more information. [ 754.865657] env[63202]: ERROR nova.compute.manager [ 754.865657] env[63202]: Traceback (most recent call last): [ 754.865657] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 754.865657] env[63202]: listener.cb(fileno) [ 754.865657] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 754.865657] env[63202]: result = function(*args, **kwargs) [ 754.865657] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 754.865657] env[63202]: return func(*args, **kwargs) [ 754.865657] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 754.865657] env[63202]: raise e [ 754.865657] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 754.865657] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 754.865657] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 754.865657] env[63202]: created_port_ids = self._update_ports_for_instance( [ 754.865657] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 754.865657] env[63202]: with excutils.save_and_reraise_exception(): [ 754.865657] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 754.865657] env[63202]: self.force_reraise() [ 754.865657] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 754.865657] env[63202]: raise self.value [ 754.865657] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 754.865657] env[63202]: updated_port = self._update_port( [ 754.865657] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 754.865657] env[63202]: _ensure_no_port_binding_failure(port) [ 754.865657] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 754.865657] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 754.866916] env[63202]: nova.exception.PortBindingFailed: Binding failed for port 0cdc9fcc-a493-46de-b498-4389665eaf90, please check neutron logs for more information. [ 754.866916] env[63202]: Removing descriptor: 16 [ 754.866916] env[63202]: ERROR nova.compute.manager [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0cdc9fcc-a493-46de-b498-4389665eaf90, please check neutron logs for more information. [ 754.866916] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] Traceback (most recent call last): [ 754.866916] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 754.866916] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] yield resources [ 754.866916] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 754.866916] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] self.driver.spawn(context, instance, image_meta, [ 754.866916] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 754.866916] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 754.866916] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 754.866916] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] vm_ref = self.build_virtual_machine(instance, [ 754.869920] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 754.869920] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] vif_infos = vmwarevif.get_vif_info(self._session, [ 754.869920] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 754.869920] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] for vif in network_info: [ 754.869920] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 754.869920] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] return self._sync_wrapper(fn, *args, **kwargs) [ 754.869920] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 754.869920] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] self.wait() [ 754.869920] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 754.869920] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] self[:] = self._gt.wait() [ 754.869920] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 754.869920] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] return self._exit_event.wait() [ 754.869920] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 754.870368] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] result = hub.switch() [ 754.870368] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 754.870368] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] return self.greenlet.switch() [ 754.870368] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 754.870368] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] result = function(*args, **kwargs) [ 754.870368] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 754.870368] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] return func(*args, **kwargs) [ 754.870368] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 754.870368] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] raise e [ 754.870368] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 754.870368] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] nwinfo = self.network_api.allocate_for_instance( [ 754.870368] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 754.870368] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] created_port_ids = self._update_ports_for_instance( [ 754.870769] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 754.870769] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] with excutils.save_and_reraise_exception(): [ 754.870769] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 754.870769] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] self.force_reraise() [ 754.870769] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 754.870769] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] raise self.value [ 754.870769] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 754.870769] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] updated_port = self._update_port( [ 754.870769] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 754.870769] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] _ensure_no_port_binding_failure(port) [ 754.870769] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 754.870769] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] raise exception.PortBindingFailed(port_id=port['id']) [ 754.871101] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] nova.exception.PortBindingFailed: Binding failed for port 0cdc9fcc-a493-46de-b498-4389665eaf90, please check neutron logs for more information. [ 754.871101] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] [ 754.871101] env[63202]: INFO nova.compute.manager [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] Terminating instance [ 754.871101] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] Acquiring lock "refresh_cache-611fead3-3d1b-41e4-9579-7ad1a1b754c3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 754.919799] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8fdaf019-1619-4801-8841-bd0e7ea87519 tempest-ServerShowV247Test-1290243861 tempest-ServerShowV247Test-1290243861-project-member] Lock "8cd0415d-dab9-4554-966c-1ea8d822deff" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.734s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.143108] env[63202]: DEBUG nova.network.neutron [req-752c2780-1e28-4b15-8d48-3889faa5795e req-cb38e587-7be5-409e-84cf-fa71661fe530 service nova] [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 755.267137] env[63202]: DEBUG nova.network.neutron [req-752c2780-1e28-4b15-8d48-3889faa5795e req-cb38e587-7be5-409e-84cf-fa71661fe530 service nova] [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 755.732887] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aaea6f9-a2fa-4b44-b317-ad5e23028ba5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.741142] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15283437-5754-448d-8ac7-f253a666ad7e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.769825] env[63202]: DEBUG oslo_concurrency.lockutils [req-752c2780-1e28-4b15-8d48-3889faa5795e req-cb38e587-7be5-409e-84cf-fa71661fe530 service nova] Releasing lock "refresh_cache-611fead3-3d1b-41e4-9579-7ad1a1b754c3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 755.770402] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] Acquired lock "refresh_cache-611fead3-3d1b-41e4-9579-7ad1a1b754c3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.770581] env[63202]: DEBUG nova.network.neutron [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 755.772150] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60909e18-c5f0-4683-85cd-c35a65aa23f8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.782187] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77514ff3-88b7-4c01-bee3-9f159c61e4ea {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.797439] env[63202]: DEBUG nova.compute.provider_tree [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 756.296792] env[63202]: DEBUG nova.network.neutron [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 756.303485] env[63202]: DEBUG nova.scheduler.client.report [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 756.582205] env[63202]: DEBUG nova.network.neutron [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.771230] env[63202]: DEBUG nova.compute.manager [req-28407b9c-886c-4297-8594-6ad307a2d24c req-41033358-9c24-470e-a066-05d5b133f8b1 service nova] [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] Received event network-vif-deleted-0cdc9fcc-a493-46de-b498-4389665eaf90 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 756.807836] env[63202]: DEBUG oslo_concurrency.lockutils [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.429s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.808383] env[63202]: DEBUG nova.compute.manager [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 756.810944] env[63202]: DEBUG oslo_concurrency.lockutils [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.861s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 756.812328] env[63202]: INFO nova.compute.claims [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 757.085152] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] Releasing lock "refresh_cache-611fead3-3d1b-41e4-9579-7ad1a1b754c3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 757.085152] env[63202]: DEBUG nova.compute.manager [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 757.085152] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 757.085314] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8269e3b0-722d-4e3d-b10e-9140e2420a4a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.096075] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07e6ae41-d575-4c4d-a343-3a3a99ca014b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.117617] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 611fead3-3d1b-41e4-9579-7ad1a1b754c3 could not be found. [ 757.117886] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 757.118056] env[63202]: INFO nova.compute.manager [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] Took 0.03 seconds to destroy the instance on the hypervisor. [ 757.118311] env[63202]: DEBUG oslo.service.loopingcall [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 757.118508] env[63202]: DEBUG nova.compute.manager [-] [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 757.118598] env[63202]: DEBUG nova.network.neutron [-] [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 757.133598] env[63202]: DEBUG nova.network.neutron [-] [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 757.317407] env[63202]: DEBUG nova.compute.utils [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 757.320671] env[63202]: DEBUG nova.compute.manager [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 757.320831] env[63202]: DEBUG nova.network.neutron [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 757.401080] env[63202]: DEBUG nova.policy [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '69493cd1ff694fe8a805f116b407ea1b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6b9be262a4894ae78a9964e5ca7b09a8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 757.635747] env[63202]: DEBUG nova.network.neutron [-] [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 757.799007] env[63202]: DEBUG nova.network.neutron [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] Successfully created port: 35cca4a9-1e33-42f9-a7d4-50c5cea28d14 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 757.821233] env[63202]: DEBUG nova.compute.manager [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 758.140374] env[63202]: INFO nova.compute.manager [-] [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] Took 1.02 seconds to deallocate network for instance. [ 758.145445] env[63202]: DEBUG nova.compute.claims [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 758.145722] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 758.175622] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6674806c-9548-4409-b72c-67feef90bbad {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.185475] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dcf5cbc-fcd9-4c92-81f6-4c809faba494 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.216851] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb69fc4a-9a65-4b29-816e-00a8332af203 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.225371] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea6ee9a4-a94e-48d4-9d0f-1ae3698d170a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.239891] env[63202]: DEBUG nova.compute.provider_tree [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 758.744083] env[63202]: DEBUG nova.scheduler.client.report [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 758.835902] env[63202]: DEBUG nova.compute.manager [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 758.838624] env[63202]: ERROR nova.compute.manager [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 35cca4a9-1e33-42f9-a7d4-50c5cea28d14, please check neutron logs for more information. [ 758.838624] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 758.838624] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 758.838624] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 758.838624] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 758.838624] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 758.838624] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 758.838624] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 758.838624] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 758.838624] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 758.838624] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 758.838624] env[63202]: ERROR nova.compute.manager raise self.value [ 758.838624] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 758.838624] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 758.838624] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 758.838624] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 758.839131] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 758.839131] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 758.839131] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 35cca4a9-1e33-42f9-a7d4-50c5cea28d14, please check neutron logs for more information. [ 758.839131] env[63202]: ERROR nova.compute.manager [ 758.839131] env[63202]: Traceback (most recent call last): [ 758.839131] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 758.839131] env[63202]: listener.cb(fileno) [ 758.839131] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 758.839131] env[63202]: result = function(*args, **kwargs) [ 758.839131] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 758.839131] env[63202]: return func(*args, **kwargs) [ 758.839131] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 758.839131] env[63202]: raise e [ 758.839131] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 758.839131] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 758.839131] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 758.839131] env[63202]: created_port_ids = self._update_ports_for_instance( [ 758.839131] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 758.839131] env[63202]: with excutils.save_and_reraise_exception(): [ 758.839131] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 758.839131] env[63202]: self.force_reraise() [ 758.839131] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 758.839131] env[63202]: raise self.value [ 758.839131] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 758.839131] env[63202]: updated_port = self._update_port( [ 758.839131] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 758.839131] env[63202]: _ensure_no_port_binding_failure(port) [ 758.839131] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 758.839131] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 758.839934] env[63202]: nova.exception.PortBindingFailed: Binding failed for port 35cca4a9-1e33-42f9-a7d4-50c5cea28d14, please check neutron logs for more information. [ 758.839934] env[63202]: Removing descriptor: 16 [ 758.840910] env[63202]: DEBUG nova.compute.manager [req-6ab26bb5-2c7b-4a4c-b187-6a3b8516de61 req-47cf2fe4-0cb3-4efe-8a8a-aec763d1bcbb service nova] [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] Received event network-changed-35cca4a9-1e33-42f9-a7d4-50c5cea28d14 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 758.841295] env[63202]: DEBUG nova.compute.manager [req-6ab26bb5-2c7b-4a4c-b187-6a3b8516de61 req-47cf2fe4-0cb3-4efe-8a8a-aec763d1bcbb service nova] [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] Refreshing instance network info cache due to event network-changed-35cca4a9-1e33-42f9-a7d4-50c5cea28d14. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 758.841295] env[63202]: DEBUG oslo_concurrency.lockutils [req-6ab26bb5-2c7b-4a4c-b187-6a3b8516de61 req-47cf2fe4-0cb3-4efe-8a8a-aec763d1bcbb service nova] Acquiring lock "refresh_cache-4cd7f7b3-d947-4745-8fd7-940076865e3b" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 758.841488] env[63202]: DEBUG oslo_concurrency.lockutils [req-6ab26bb5-2c7b-4a4c-b187-6a3b8516de61 req-47cf2fe4-0cb3-4efe-8a8a-aec763d1bcbb service nova] Acquired lock "refresh_cache-4cd7f7b3-d947-4745-8fd7-940076865e3b" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.841616] env[63202]: DEBUG nova.network.neutron [req-6ab26bb5-2c7b-4a4c-b187-6a3b8516de61 req-47cf2fe4-0cb3-4efe-8a8a-aec763d1bcbb service nova] [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] Refreshing network info cache for port 35cca4a9-1e33-42f9-a7d4-50c5cea28d14 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 758.877369] env[63202]: DEBUG nova.virt.hardware [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 758.877790] env[63202]: DEBUG nova.virt.hardware [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 758.878051] env[63202]: DEBUG nova.virt.hardware [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 758.878325] env[63202]: DEBUG nova.virt.hardware [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 758.878566] env[63202]: DEBUG nova.virt.hardware [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 758.878849] env[63202]: DEBUG nova.virt.hardware [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 758.879189] env[63202]: DEBUG nova.virt.hardware [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 758.879450] env[63202]: DEBUG nova.virt.hardware [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 758.880028] env[63202]: DEBUG nova.virt.hardware [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 758.880356] env[63202]: DEBUG nova.virt.hardware [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 758.880807] env[63202]: DEBUG nova.virt.hardware [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 758.882142] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5accdd97-7539-4e51-b948-76d0018ecb61 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.893464] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bfbac72-5b82-4dc1-a631-435677a7594a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.914196] env[63202]: ERROR nova.compute.manager [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 35cca4a9-1e33-42f9-a7d4-50c5cea28d14, please check neutron logs for more information. [ 758.914196] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] Traceback (most recent call last): [ 758.914196] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 758.914196] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] yield resources [ 758.914196] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 758.914196] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] self.driver.spawn(context, instance, image_meta, [ 758.914196] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 758.914196] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 758.914196] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 758.914196] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] vm_ref = self.build_virtual_machine(instance, [ 758.914196] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 758.914624] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] vif_infos = vmwarevif.get_vif_info(self._session, [ 758.914624] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 758.914624] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] for vif in network_info: [ 758.914624] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 758.914624] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] return self._sync_wrapper(fn, *args, **kwargs) [ 758.914624] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 758.914624] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] self.wait() [ 758.914624] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 758.914624] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] self[:] = self._gt.wait() [ 758.914624] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 758.914624] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] return self._exit_event.wait() [ 758.914624] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 758.914624] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] current.throw(*self._exc) [ 758.914955] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 758.914955] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] result = function(*args, **kwargs) [ 758.914955] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 758.914955] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] return func(*args, **kwargs) [ 758.914955] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 758.914955] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] raise e [ 758.914955] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 758.914955] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] nwinfo = self.network_api.allocate_for_instance( [ 758.914955] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 758.914955] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] created_port_ids = self._update_ports_for_instance( [ 758.914955] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 758.914955] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] with excutils.save_and_reraise_exception(): [ 758.914955] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 758.915285] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] self.force_reraise() [ 758.915285] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 758.915285] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] raise self.value [ 758.915285] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 758.915285] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] updated_port = self._update_port( [ 758.915285] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 758.915285] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] _ensure_no_port_binding_failure(port) [ 758.915285] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 758.915285] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] raise exception.PortBindingFailed(port_id=port['id']) [ 758.915285] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] nova.exception.PortBindingFailed: Binding failed for port 35cca4a9-1e33-42f9-a7d4-50c5cea28d14, please check neutron logs for more information. [ 758.915285] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] [ 758.915285] env[63202]: INFO nova.compute.manager [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] Terminating instance [ 758.921441] env[63202]: DEBUG oslo_concurrency.lockutils [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Acquiring lock "refresh_cache-4cd7f7b3-d947-4745-8fd7-940076865e3b" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.250387] env[63202]: DEBUG oslo_concurrency.lockutils [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.439s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.251423] env[63202]: DEBUG nova.compute.manager [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 759.254024] env[63202]: DEBUG oslo_concurrency.lockutils [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.336s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 759.255604] env[63202]: INFO nova.compute.claims [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 759.377781] env[63202]: DEBUG nova.network.neutron [req-6ab26bb5-2c7b-4a4c-b187-6a3b8516de61 req-47cf2fe4-0cb3-4efe-8a8a-aec763d1bcbb service nova] [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 759.452119] env[63202]: DEBUG nova.network.neutron [req-6ab26bb5-2c7b-4a4c-b187-6a3b8516de61 req-47cf2fe4-0cb3-4efe-8a8a-aec763d1bcbb service nova] [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.760188] env[63202]: DEBUG nova.compute.utils [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 759.763668] env[63202]: DEBUG nova.compute.manager [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 759.763838] env[63202]: DEBUG nova.network.neutron [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 759.822797] env[63202]: DEBUG nova.policy [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '69493cd1ff694fe8a805f116b407ea1b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6b9be262a4894ae78a9964e5ca7b09a8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 759.957332] env[63202]: DEBUG oslo_concurrency.lockutils [req-6ab26bb5-2c7b-4a4c-b187-6a3b8516de61 req-47cf2fe4-0cb3-4efe-8a8a-aec763d1bcbb service nova] Releasing lock "refresh_cache-4cd7f7b3-d947-4745-8fd7-940076865e3b" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 759.958770] env[63202]: DEBUG oslo_concurrency.lockutils [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Acquired lock "refresh_cache-4cd7f7b3-d947-4745-8fd7-940076865e3b" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.958770] env[63202]: DEBUG nova.network.neutron [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 760.244747] env[63202]: DEBUG nova.network.neutron [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] Successfully created port: a2c4847c-da1f-4c90-87f0-4678eeaf1504 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 760.264557] env[63202]: DEBUG nova.compute.manager [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 760.392151] env[63202]: DEBUG oslo_concurrency.lockutils [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "d674db7a-8cf0-46f8-b676-fced0767c105" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.392261] env[63202]: DEBUG oslo_concurrency.lockutils [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "d674db7a-8cf0-46f8-b676-fced0767c105" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.482135] env[63202]: DEBUG nova.network.neutron [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 760.565322] env[63202]: DEBUG nova.network.neutron [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.628964] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b8de93c-0a82-410f-a0cb-20a480cb1b1d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.636370] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-add2af68-fe27-4e16-9d6c-e84b6d071f6e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.667433] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01c1f81e-2067-4877-9b69-190ac4757931 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.675382] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b73f432-7872-4521-9459-babfd4c01b16 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.688798] env[63202]: DEBUG nova.compute.provider_tree [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 760.896677] env[63202]: DEBUG nova.compute.manager [req-18c8345a-3502-4b83-9200-4afde945cdc6 req-db4b3bcb-6cb8-471f-bbfa-0cfa00122465 service nova] [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] Received event network-vif-deleted-35cca4a9-1e33-42f9-a7d4-50c5cea28d14 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 761.068199] env[63202]: DEBUG oslo_concurrency.lockutils [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Releasing lock "refresh_cache-4cd7f7b3-d947-4745-8fd7-940076865e3b" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 761.068603] env[63202]: DEBUG nova.compute.manager [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 761.068799] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 761.069512] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1384075b-d2ed-4ca8-a06a-a56f82421372 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.079832] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38dcc142-fac2-4233-922c-82c3cd844fa4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.105072] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4cd7f7b3-d947-4745-8fd7-940076865e3b could not be found. [ 761.105072] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 761.105072] env[63202]: INFO nova.compute.manager [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] Took 0.04 seconds to destroy the instance on the hypervisor. [ 761.106034] env[63202]: DEBUG oslo.service.loopingcall [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 761.106034] env[63202]: DEBUG nova.compute.manager [-] [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 761.106034] env[63202]: DEBUG nova.network.neutron [-] [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 761.124036] env[63202]: DEBUG nova.network.neutron [-] [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 761.191841] env[63202]: DEBUG nova.scheduler.client.report [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 761.278413] env[63202]: DEBUG nova.compute.manager [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 761.311883] env[63202]: DEBUG nova.virt.hardware [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 761.311973] env[63202]: DEBUG nova.virt.hardware [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 761.312350] env[63202]: DEBUG nova.virt.hardware [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 761.312350] env[63202]: DEBUG nova.virt.hardware [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 761.312562] env[63202]: DEBUG nova.virt.hardware [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 761.312718] env[63202]: DEBUG nova.virt.hardware [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 761.312985] env[63202]: DEBUG nova.virt.hardware [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 761.313561] env[63202]: DEBUG nova.virt.hardware [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 761.313561] env[63202]: DEBUG nova.virt.hardware [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 761.313561] env[63202]: DEBUG nova.virt.hardware [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 761.313926] env[63202]: DEBUG nova.virt.hardware [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 761.314665] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d7599ca-2a47-433f-9284-05cd33227a21 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.322695] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0678240-3a23-4899-9797-c7f55d566f45 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.375444] env[63202]: ERROR nova.compute.manager [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a2c4847c-da1f-4c90-87f0-4678eeaf1504, please check neutron logs for more information. [ 761.375444] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 761.375444] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 761.375444] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 761.375444] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 761.375444] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 761.375444] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 761.375444] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 761.375444] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 761.375444] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 761.375444] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 761.375444] env[63202]: ERROR nova.compute.manager raise self.value [ 761.375444] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 761.375444] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 761.375444] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 761.375444] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 761.375913] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 761.375913] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 761.375913] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a2c4847c-da1f-4c90-87f0-4678eeaf1504, please check neutron logs for more information. [ 761.375913] env[63202]: ERROR nova.compute.manager [ 761.375913] env[63202]: Traceback (most recent call last): [ 761.375913] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 761.375913] env[63202]: listener.cb(fileno) [ 761.375913] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 761.375913] env[63202]: result = function(*args, **kwargs) [ 761.375913] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 761.375913] env[63202]: return func(*args, **kwargs) [ 761.375913] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 761.375913] env[63202]: raise e [ 761.375913] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 761.375913] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 761.375913] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 761.375913] env[63202]: created_port_ids = self._update_ports_for_instance( [ 761.375913] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 761.375913] env[63202]: with excutils.save_and_reraise_exception(): [ 761.375913] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 761.375913] env[63202]: self.force_reraise() [ 761.375913] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 761.375913] env[63202]: raise self.value [ 761.375913] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 761.375913] env[63202]: updated_port = self._update_port( [ 761.375913] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 761.375913] env[63202]: _ensure_no_port_binding_failure(port) [ 761.375913] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 761.375913] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 761.376728] env[63202]: nova.exception.PortBindingFailed: Binding failed for port a2c4847c-da1f-4c90-87f0-4678eeaf1504, please check neutron logs for more information. [ 761.376728] env[63202]: Removing descriptor: 16 [ 761.376728] env[63202]: ERROR nova.compute.manager [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a2c4847c-da1f-4c90-87f0-4678eeaf1504, please check neutron logs for more information. [ 761.376728] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] Traceback (most recent call last): [ 761.376728] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 761.376728] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] yield resources [ 761.376728] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 761.376728] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] self.driver.spawn(context, instance, image_meta, [ 761.376728] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 761.376728] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] self._vmops.spawn(context, instance, image_meta, injected_files, [ 761.376728] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 761.376728] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] vm_ref = self.build_virtual_machine(instance, [ 761.377080] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 761.377080] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] vif_infos = vmwarevif.get_vif_info(self._session, [ 761.377080] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 761.377080] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] for vif in network_info: [ 761.377080] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 761.377080] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] return self._sync_wrapper(fn, *args, **kwargs) [ 761.377080] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 761.377080] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] self.wait() [ 761.377080] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 761.377080] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] self[:] = self._gt.wait() [ 761.377080] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 761.377080] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] return self._exit_event.wait() [ 761.377080] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 761.377457] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] result = hub.switch() [ 761.377457] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 761.377457] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] return self.greenlet.switch() [ 761.377457] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 761.377457] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] result = function(*args, **kwargs) [ 761.377457] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 761.377457] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] return func(*args, **kwargs) [ 761.377457] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 761.377457] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] raise e [ 761.377457] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 761.377457] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] nwinfo = self.network_api.allocate_for_instance( [ 761.377457] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 761.377457] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] created_port_ids = self._update_ports_for_instance( [ 761.377894] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 761.377894] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] with excutils.save_and_reraise_exception(): [ 761.377894] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 761.377894] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] self.force_reraise() [ 761.377894] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 761.377894] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] raise self.value [ 761.377894] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 761.377894] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] updated_port = self._update_port( [ 761.377894] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 761.377894] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] _ensure_no_port_binding_failure(port) [ 761.377894] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 761.377894] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] raise exception.PortBindingFailed(port_id=port['id']) [ 761.378279] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] nova.exception.PortBindingFailed: Binding failed for port a2c4847c-da1f-4c90-87f0-4678eeaf1504, please check neutron logs for more information. [ 761.378279] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] [ 761.378279] env[63202]: INFO nova.compute.manager [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] Terminating instance [ 761.378793] env[63202]: DEBUG oslo_concurrency.lockutils [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Acquiring lock "refresh_cache-ba2d047d-d30f-4ba0-bcfb-787c5a3ae516" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.378957] env[63202]: DEBUG oslo_concurrency.lockutils [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Acquired lock "refresh_cache-ba2d047d-d30f-4ba0-bcfb-787c5a3ae516" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.379132] env[63202]: DEBUG nova.network.neutron [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 761.626902] env[63202]: DEBUG nova.network.neutron [-] [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.697397] env[63202]: DEBUG oslo_concurrency.lockutils [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.443s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.698402] env[63202]: DEBUG nova.compute.manager [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 761.701065] env[63202]: DEBUG oslo_concurrency.lockutils [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.758s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.702613] env[63202]: INFO nova.compute.claims [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 761.897179] env[63202]: DEBUG nova.network.neutron [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 762.005792] env[63202]: DEBUG nova.network.neutron [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.055798] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 762.056124] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 762.056308] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Starting heal instance info cache {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 762.056454] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Rebuilding the list of instances to heal {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 762.129625] env[63202]: INFO nova.compute.manager [-] [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] Took 1.02 seconds to deallocate network for instance. [ 762.132380] env[63202]: DEBUG nova.compute.claims [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 762.132629] env[63202]: DEBUG oslo_concurrency.lockutils [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.212130] env[63202]: DEBUG nova.compute.utils [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 762.215238] env[63202]: DEBUG nova.compute.manager [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 762.215415] env[63202]: DEBUG nova.network.neutron [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 762.294624] env[63202]: DEBUG nova.policy [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1a25e967c31e41debbe21da7e5dd463e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9200b0e7234d4c09826bd5166b0c05fe', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 762.511393] env[63202]: DEBUG oslo_concurrency.lockutils [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Releasing lock "refresh_cache-ba2d047d-d30f-4ba0-bcfb-787c5a3ae516" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.511393] env[63202]: DEBUG nova.compute.manager [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 762.511393] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 762.511393] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ce26b3b4-82ae-4d6b-a0d0-b1b873135ff6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.521094] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b332bd1-4ae1-4d9c-ac4b-35be20c542db {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.542668] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ba2d047d-d30f-4ba0-bcfb-787c5a3ae516 could not be found. [ 762.542901] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 762.543097] env[63202]: INFO nova.compute.manager [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] Took 0.03 seconds to destroy the instance on the hypervisor. [ 762.543335] env[63202]: DEBUG oslo.service.loopingcall [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 762.543554] env[63202]: DEBUG nova.compute.manager [-] [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 762.543645] env[63202]: DEBUG nova.network.neutron [-] [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 762.558564] env[63202]: DEBUG nova.network.neutron [-] [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 762.560918] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] Skipping network cache update for instance because it is Building. {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 762.561124] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] Skipping network cache update for instance because it is Building. {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 762.561281] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] Skipping network cache update for instance because it is Building. {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 762.561428] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] Skipping network cache update for instance because it is Building. {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 762.561570] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] Skipping network cache update for instance because it is Building. {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 762.561714] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Didn't find any instances for network info cache update. {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 762.562130] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 762.562333] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 762.562514] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 762.562695] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 762.562868] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 762.563054] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 762.563210] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63202) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 762.563391] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager.update_available_resource {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 762.678302] env[63202]: DEBUG nova.network.neutron [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] Successfully created port: cc4ef3ae-fbb2-4f70-b9e7-103b29e5b9f6 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 762.716603] env[63202]: DEBUG nova.compute.manager [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 762.929622] env[63202]: DEBUG nova.compute.manager [req-8b0bf497-2a2b-483b-ab02-9b42ffc9ab30 req-77d73de8-6320-450d-b5b6-36ef0d05d78b service nova] [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] Received event network-changed-a2c4847c-da1f-4c90-87f0-4678eeaf1504 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 762.929804] env[63202]: DEBUG nova.compute.manager [req-8b0bf497-2a2b-483b-ab02-9b42ffc9ab30 req-77d73de8-6320-450d-b5b6-36ef0d05d78b service nova] [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] Refreshing instance network info cache due to event network-changed-a2c4847c-da1f-4c90-87f0-4678eeaf1504. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 762.930012] env[63202]: DEBUG oslo_concurrency.lockutils [req-8b0bf497-2a2b-483b-ab02-9b42ffc9ab30 req-77d73de8-6320-450d-b5b6-36ef0d05d78b service nova] Acquiring lock "refresh_cache-ba2d047d-d30f-4ba0-bcfb-787c5a3ae516" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.930149] env[63202]: DEBUG oslo_concurrency.lockutils [req-8b0bf497-2a2b-483b-ab02-9b42ffc9ab30 req-77d73de8-6320-450d-b5b6-36ef0d05d78b service nova] Acquired lock "refresh_cache-ba2d047d-d30f-4ba0-bcfb-787c5a3ae516" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.932022] env[63202]: DEBUG nova.network.neutron [req-8b0bf497-2a2b-483b-ab02-9b42ffc9ab30 req-77d73de8-6320-450d-b5b6-36ef0d05d78b service nova] [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] Refreshing network info cache for port a2c4847c-da1f-4c90-87f0-4678eeaf1504 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 763.035352] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44cd1d67-0d4f-4119-a933-8ff3472e99cb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.042847] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a57b010-bf02-4314-9170-370e2b8b8376 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.073436] env[63202]: DEBUG nova.network.neutron [-] [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.075171] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 763.076499] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c2cdf04-5fca-46eb-8b46-1c2ea56bbf18 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.090568] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e97b80c5-6ad1-4ba7-b772-f55bfa1fa3ad {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.104370] env[63202]: DEBUG nova.compute.provider_tree [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 763.449533] env[63202]: DEBUG nova.network.neutron [req-8b0bf497-2a2b-483b-ab02-9b42ffc9ab30 req-77d73de8-6320-450d-b5b6-36ef0d05d78b service nova] [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 763.530131] env[63202]: DEBUG nova.network.neutron [req-8b0bf497-2a2b-483b-ab02-9b42ffc9ab30 req-77d73de8-6320-450d-b5b6-36ef0d05d78b service nova] [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.585534] env[63202]: INFO nova.compute.manager [-] [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] Took 1.04 seconds to deallocate network for instance. [ 763.588718] env[63202]: DEBUG nova.compute.claims [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 763.588788] env[63202]: DEBUG oslo_concurrency.lockutils [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 763.606886] env[63202]: DEBUG nova.scheduler.client.report [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 763.729204] env[63202]: DEBUG nova.compute.manager [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 763.753373] env[63202]: DEBUG nova.virt.hardware [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 763.753621] env[63202]: DEBUG nova.virt.hardware [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 763.754180] env[63202]: DEBUG nova.virt.hardware [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 763.754180] env[63202]: DEBUG nova.virt.hardware [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 763.754180] env[63202]: DEBUG nova.virt.hardware [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 763.754327] env[63202]: DEBUG nova.virt.hardware [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 763.754408] env[63202]: DEBUG nova.virt.hardware [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 763.754556] env[63202]: DEBUG nova.virt.hardware [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 763.754719] env[63202]: DEBUG nova.virt.hardware [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 763.754879] env[63202]: DEBUG nova.virt.hardware [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 763.755061] env[63202]: DEBUG nova.virt.hardware [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 763.755900] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d2c61ed-895f-4f76-b109-facefd8220b6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.763900] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4caa22af-aed7-4821-b33f-f5f4a9f0b2a2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.928131] env[63202]: ERROR nova.compute.manager [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port cc4ef3ae-fbb2-4f70-b9e7-103b29e5b9f6, please check neutron logs for more information. [ 763.928131] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 763.928131] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 763.928131] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 763.928131] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 763.928131] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 763.928131] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 763.928131] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 763.928131] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 763.928131] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 763.928131] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 763.928131] env[63202]: ERROR nova.compute.manager raise self.value [ 763.928131] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 763.928131] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 763.928131] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 763.928131] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 763.928641] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 763.928641] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 763.928641] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port cc4ef3ae-fbb2-4f70-b9e7-103b29e5b9f6, please check neutron logs for more information. [ 763.928641] env[63202]: ERROR nova.compute.manager [ 763.928641] env[63202]: Traceback (most recent call last): [ 763.928641] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 763.928641] env[63202]: listener.cb(fileno) [ 763.928641] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 763.928641] env[63202]: result = function(*args, **kwargs) [ 763.928641] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 763.928641] env[63202]: return func(*args, **kwargs) [ 763.928641] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 763.928641] env[63202]: raise e [ 763.928641] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 763.928641] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 763.928641] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 763.928641] env[63202]: created_port_ids = self._update_ports_for_instance( [ 763.928641] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 763.928641] env[63202]: with excutils.save_and_reraise_exception(): [ 763.928641] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 763.928641] env[63202]: self.force_reraise() [ 763.928641] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 763.928641] env[63202]: raise self.value [ 763.928641] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 763.928641] env[63202]: updated_port = self._update_port( [ 763.928641] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 763.928641] env[63202]: _ensure_no_port_binding_failure(port) [ 763.928641] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 763.928641] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 763.929566] env[63202]: nova.exception.PortBindingFailed: Binding failed for port cc4ef3ae-fbb2-4f70-b9e7-103b29e5b9f6, please check neutron logs for more information. [ 763.929566] env[63202]: Removing descriptor: 16 [ 763.929566] env[63202]: ERROR nova.compute.manager [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port cc4ef3ae-fbb2-4f70-b9e7-103b29e5b9f6, please check neutron logs for more information. [ 763.929566] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] Traceback (most recent call last): [ 763.929566] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 763.929566] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] yield resources [ 763.929566] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 763.929566] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] self.driver.spawn(context, instance, image_meta, [ 763.929566] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 763.929566] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 763.929566] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 763.929566] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] vm_ref = self.build_virtual_machine(instance, [ 763.929964] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 763.929964] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] vif_infos = vmwarevif.get_vif_info(self._session, [ 763.929964] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 763.929964] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] for vif in network_info: [ 763.929964] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 763.929964] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] return self._sync_wrapper(fn, *args, **kwargs) [ 763.929964] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 763.929964] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] self.wait() [ 763.929964] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 763.929964] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] self[:] = self._gt.wait() [ 763.929964] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 763.929964] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] return self._exit_event.wait() [ 763.929964] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 763.930358] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] result = hub.switch() [ 763.930358] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 763.930358] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] return self.greenlet.switch() [ 763.930358] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 763.930358] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] result = function(*args, **kwargs) [ 763.930358] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 763.930358] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] return func(*args, **kwargs) [ 763.930358] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 763.930358] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] raise e [ 763.930358] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 763.930358] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] nwinfo = self.network_api.allocate_for_instance( [ 763.930358] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 763.930358] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] created_port_ids = self._update_ports_for_instance( [ 763.930746] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 763.930746] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] with excutils.save_and_reraise_exception(): [ 763.930746] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 763.930746] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] self.force_reraise() [ 763.930746] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 763.930746] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] raise self.value [ 763.930746] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 763.930746] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] updated_port = self._update_port( [ 763.930746] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 763.930746] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] _ensure_no_port_binding_failure(port) [ 763.930746] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 763.930746] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] raise exception.PortBindingFailed(port_id=port['id']) [ 763.931123] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] nova.exception.PortBindingFailed: Binding failed for port cc4ef3ae-fbb2-4f70-b9e7-103b29e5b9f6, please check neutron logs for more information. [ 763.931123] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] [ 763.931123] env[63202]: INFO nova.compute.manager [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] Terminating instance [ 763.931411] env[63202]: DEBUG oslo_concurrency.lockutils [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Acquiring lock "refresh_cache-1ac8e024-65d1-4250-a946-858aaff852c8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 763.931593] env[63202]: DEBUG oslo_concurrency.lockutils [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Acquired lock "refresh_cache-1ac8e024-65d1-4250-a946-858aaff852c8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.931757] env[63202]: DEBUG nova.network.neutron [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 764.034940] env[63202]: DEBUG oslo_concurrency.lockutils [req-8b0bf497-2a2b-483b-ab02-9b42ffc9ab30 req-77d73de8-6320-450d-b5b6-36ef0d05d78b service nova] Releasing lock "refresh_cache-ba2d047d-d30f-4ba0-bcfb-787c5a3ae516" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.035175] env[63202]: DEBUG nova.compute.manager [req-8b0bf497-2a2b-483b-ab02-9b42ffc9ab30 req-77d73de8-6320-450d-b5b6-36ef0d05d78b service nova] [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] Received event network-vif-deleted-a2c4847c-da1f-4c90-87f0-4678eeaf1504 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 764.112716] env[63202]: DEBUG oslo_concurrency.lockutils [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.412s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.113241] env[63202]: DEBUG nova.compute.manager [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 764.115731] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.104s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.121019] env[63202]: INFO nova.compute.claims [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 764.450653] env[63202]: DEBUG nova.network.neutron [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 764.531614] env[63202]: DEBUG nova.network.neutron [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.624890] env[63202]: DEBUG nova.compute.utils [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 764.626562] env[63202]: DEBUG nova.compute.manager [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 764.626787] env[63202]: DEBUG nova.network.neutron [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 764.690034] env[63202]: DEBUG nova.policy [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1a25e967c31e41debbe21da7e5dd463e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9200b0e7234d4c09826bd5166b0c05fe', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 764.951054] env[63202]: DEBUG nova.network.neutron [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] Successfully created port: 3c2c28dc-66c3-479c-a3a2-2d1ed7c3ae4d {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 764.967405] env[63202]: DEBUG nova.compute.manager [req-b2898453-071e-4ef6-b58d-4351b0466c0b req-8fb0c56e-836e-4b2e-9cda-ecbda5188667 service nova] [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] Received event network-changed-cc4ef3ae-fbb2-4f70-b9e7-103b29e5b9f6 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 764.967588] env[63202]: DEBUG nova.compute.manager [req-b2898453-071e-4ef6-b58d-4351b0466c0b req-8fb0c56e-836e-4b2e-9cda-ecbda5188667 service nova] [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] Refreshing instance network info cache due to event network-changed-cc4ef3ae-fbb2-4f70-b9e7-103b29e5b9f6. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 764.967795] env[63202]: DEBUG oslo_concurrency.lockutils [req-b2898453-071e-4ef6-b58d-4351b0466c0b req-8fb0c56e-836e-4b2e-9cda-ecbda5188667 service nova] Acquiring lock "refresh_cache-1ac8e024-65d1-4250-a946-858aaff852c8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.033751] env[63202]: DEBUG oslo_concurrency.lockutils [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Releasing lock "refresh_cache-1ac8e024-65d1-4250-a946-858aaff852c8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 765.034214] env[63202]: DEBUG nova.compute.manager [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 765.034405] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 765.034698] env[63202]: DEBUG oslo_concurrency.lockutils [req-b2898453-071e-4ef6-b58d-4351b0466c0b req-8fb0c56e-836e-4b2e-9cda-ecbda5188667 service nova] Acquired lock "refresh_cache-1ac8e024-65d1-4250-a946-858aaff852c8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.034864] env[63202]: DEBUG nova.network.neutron [req-b2898453-071e-4ef6-b58d-4351b0466c0b req-8fb0c56e-836e-4b2e-9cda-ecbda5188667 service nova] [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] Refreshing network info cache for port cc4ef3ae-fbb2-4f70-b9e7-103b29e5b9f6 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 765.035873] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a74aa009-28af-4c4f-94bb-fde13227aa58 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.045601] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-facd9eb7-c050-4eff-96f2-bb145f7edbd3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.067201] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1ac8e024-65d1-4250-a946-858aaff852c8 could not be found. [ 765.067415] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 765.067589] env[63202]: INFO nova.compute.manager [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] Took 0.03 seconds to destroy the instance on the hypervisor. [ 765.067816] env[63202]: DEBUG oslo.service.loopingcall [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 765.068014] env[63202]: DEBUG nova.compute.manager [-] [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 765.068107] env[63202]: DEBUG nova.network.neutron [-] [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 765.085116] env[63202]: DEBUG nova.network.neutron [-] [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 765.130723] env[63202]: DEBUG nova.compute.manager [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 765.407468] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5228be02-1473-4692-9601-e17b4c547643 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.418033] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0f01a4a-5b7f-4feb-8079-430daae9428f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.448467] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ef24423-06da-40da-9414-35fe0396f66e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.458195] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dae939cd-11e7-42a9-bc85-e9b91a467d9b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.468599] env[63202]: DEBUG nova.compute.provider_tree [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 765.555949] env[63202]: DEBUG nova.network.neutron [req-b2898453-071e-4ef6-b58d-4351b0466c0b req-8fb0c56e-836e-4b2e-9cda-ecbda5188667 service nova] [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 765.589020] env[63202]: DEBUG nova.network.neutron [-] [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.627699] env[63202]: DEBUG nova.network.neutron [req-b2898453-071e-4ef6-b58d-4351b0466c0b req-8fb0c56e-836e-4b2e-9cda-ecbda5188667 service nova] [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.971373] env[63202]: DEBUG nova.scheduler.client.report [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 765.999659] env[63202]: ERROR nova.compute.manager [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3c2c28dc-66c3-479c-a3a2-2d1ed7c3ae4d, please check neutron logs for more information. [ 765.999659] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 765.999659] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 765.999659] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 765.999659] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 765.999659] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 765.999659] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 765.999659] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 765.999659] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 765.999659] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 765.999659] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 765.999659] env[63202]: ERROR nova.compute.manager raise self.value [ 765.999659] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 765.999659] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 765.999659] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 765.999659] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 766.000214] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 766.000214] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 766.000214] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3c2c28dc-66c3-479c-a3a2-2d1ed7c3ae4d, please check neutron logs for more information. [ 766.000214] env[63202]: ERROR nova.compute.manager [ 766.000214] env[63202]: Traceback (most recent call last): [ 766.000214] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 766.000214] env[63202]: listener.cb(fileno) [ 766.000214] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 766.000214] env[63202]: result = function(*args, **kwargs) [ 766.000214] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 766.000214] env[63202]: return func(*args, **kwargs) [ 766.000214] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 766.000214] env[63202]: raise e [ 766.000214] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 766.000214] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 766.000214] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 766.000214] env[63202]: created_port_ids = self._update_ports_for_instance( [ 766.000214] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 766.000214] env[63202]: with excutils.save_and_reraise_exception(): [ 766.000214] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 766.000214] env[63202]: self.force_reraise() [ 766.000214] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 766.000214] env[63202]: raise self.value [ 766.000214] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 766.000214] env[63202]: updated_port = self._update_port( [ 766.000214] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 766.000214] env[63202]: _ensure_no_port_binding_failure(port) [ 766.000214] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 766.000214] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 766.001092] env[63202]: nova.exception.PortBindingFailed: Binding failed for port 3c2c28dc-66c3-479c-a3a2-2d1ed7c3ae4d, please check neutron logs for more information. [ 766.001092] env[63202]: Removing descriptor: 16 [ 766.090473] env[63202]: INFO nova.compute.manager [-] [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] Took 1.02 seconds to deallocate network for instance. [ 766.093319] env[63202]: DEBUG nova.compute.claims [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 766.093619] env[63202]: DEBUG oslo_concurrency.lockutils [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 766.130726] env[63202]: DEBUG oslo_concurrency.lockutils [req-b2898453-071e-4ef6-b58d-4351b0466c0b req-8fb0c56e-836e-4b2e-9cda-ecbda5188667 service nova] Releasing lock "refresh_cache-1ac8e024-65d1-4250-a946-858aaff852c8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 766.130978] env[63202]: DEBUG nova.compute.manager [req-b2898453-071e-4ef6-b58d-4351b0466c0b req-8fb0c56e-836e-4b2e-9cda-ecbda5188667 service nova] [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] Received event network-vif-deleted-cc4ef3ae-fbb2-4f70-b9e7-103b29e5b9f6 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 766.142357] env[63202]: DEBUG nova.compute.manager [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 766.167784] env[63202]: DEBUG nova.virt.hardware [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 766.168029] env[63202]: DEBUG nova.virt.hardware [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 766.168187] env[63202]: DEBUG nova.virt.hardware [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 766.168364] env[63202]: DEBUG nova.virt.hardware [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 766.168502] env[63202]: DEBUG nova.virt.hardware [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 766.168642] env[63202]: DEBUG nova.virt.hardware [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 766.168868] env[63202]: DEBUG nova.virt.hardware [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 766.169044] env[63202]: DEBUG nova.virt.hardware [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 766.169240] env[63202]: DEBUG nova.virt.hardware [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 766.169450] env[63202]: DEBUG nova.virt.hardware [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 766.169625] env[63202]: DEBUG nova.virt.hardware [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 766.170487] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26f52a9a-5432-403a-9cfc-8d10952cb10e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.178748] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fc19992-7b0a-429f-a9a2-ad2d1eb23cc9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.192392] env[63202]: ERROR nova.compute.manager [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3c2c28dc-66c3-479c-a3a2-2d1ed7c3ae4d, please check neutron logs for more information. [ 766.192392] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] Traceback (most recent call last): [ 766.192392] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 766.192392] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] yield resources [ 766.192392] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 766.192392] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] self.driver.spawn(context, instance, image_meta, [ 766.192392] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 766.192392] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] self._vmops.spawn(context, instance, image_meta, injected_files, [ 766.192392] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 766.192392] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] vm_ref = self.build_virtual_machine(instance, [ 766.192392] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 766.192735] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] vif_infos = vmwarevif.get_vif_info(self._session, [ 766.192735] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 766.192735] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] for vif in network_info: [ 766.192735] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 766.192735] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] return self._sync_wrapper(fn, *args, **kwargs) [ 766.192735] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 766.192735] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] self.wait() [ 766.192735] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 766.192735] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] self[:] = self._gt.wait() [ 766.192735] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 766.192735] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] return self._exit_event.wait() [ 766.192735] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 766.192735] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] current.throw(*self._exc) [ 766.193094] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 766.193094] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] result = function(*args, **kwargs) [ 766.193094] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 766.193094] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] return func(*args, **kwargs) [ 766.193094] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 766.193094] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] raise e [ 766.193094] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 766.193094] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] nwinfo = self.network_api.allocate_for_instance( [ 766.193094] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 766.193094] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] created_port_ids = self._update_ports_for_instance( [ 766.193094] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 766.193094] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] with excutils.save_and_reraise_exception(): [ 766.193094] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 766.193450] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] self.force_reraise() [ 766.193450] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 766.193450] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] raise self.value [ 766.193450] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 766.193450] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] updated_port = self._update_port( [ 766.193450] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 766.193450] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] _ensure_no_port_binding_failure(port) [ 766.193450] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 766.193450] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] raise exception.PortBindingFailed(port_id=port['id']) [ 766.193450] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] nova.exception.PortBindingFailed: Binding failed for port 3c2c28dc-66c3-479c-a3a2-2d1ed7c3ae4d, please check neutron logs for more information. [ 766.193450] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] [ 766.193450] env[63202]: INFO nova.compute.manager [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] Terminating instance [ 766.194605] env[63202]: DEBUG oslo_concurrency.lockutils [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Acquiring lock "refresh_cache-3080314c-938e-4c27-bffd-547bdc7e6e38" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 766.194759] env[63202]: DEBUG oslo_concurrency.lockutils [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Acquired lock "refresh_cache-3080314c-938e-4c27-bffd-547bdc7e6e38" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.194917] env[63202]: DEBUG nova.network.neutron [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 766.476431] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.361s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 766.476824] env[63202]: DEBUG nova.compute.manager [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 766.479859] env[63202]: DEBUG oslo_concurrency.lockutils [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.529s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 766.481241] env[63202]: INFO nova.compute.claims [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: fea91da2-186a-44b6-8e20-535e3b123890] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 766.722510] env[63202]: DEBUG nova.network.neutron [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 766.801529] env[63202]: DEBUG nova.network.neutron [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.985452] env[63202]: DEBUG nova.compute.utils [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 766.990796] env[63202]: DEBUG nova.compute.manager [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 766.990964] env[63202]: DEBUG nova.network.neutron [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 766.993544] env[63202]: DEBUG nova.compute.manager [req-9719aba8-582e-414c-b14e-dfe1481eb9b7 req-c5127086-6e7a-4690-a6a3-a436c187ae08 service nova] [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] Received event network-changed-3c2c28dc-66c3-479c-a3a2-2d1ed7c3ae4d {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 766.993544] env[63202]: DEBUG nova.compute.manager [req-9719aba8-582e-414c-b14e-dfe1481eb9b7 req-c5127086-6e7a-4690-a6a3-a436c187ae08 service nova] [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] Refreshing instance network info cache due to event network-changed-3c2c28dc-66c3-479c-a3a2-2d1ed7c3ae4d. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 766.993689] env[63202]: DEBUG oslo_concurrency.lockutils [req-9719aba8-582e-414c-b14e-dfe1481eb9b7 req-c5127086-6e7a-4690-a6a3-a436c187ae08 service nova] Acquiring lock "refresh_cache-3080314c-938e-4c27-bffd-547bdc7e6e38" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 767.030141] env[63202]: DEBUG nova.policy [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e6498e4c711949d89248b83cdbb92939', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4fdb4fdcc5134acd86354defcda8c6d8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 767.281931] env[63202]: DEBUG nova.network.neutron [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Successfully created port: 52847cc7-89c1-4996-92bd-ec6cefb47d70 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 767.305414] env[63202]: DEBUG oslo_concurrency.lockutils [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Releasing lock "refresh_cache-3080314c-938e-4c27-bffd-547bdc7e6e38" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 767.305794] env[63202]: DEBUG nova.compute.manager [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 767.305976] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 767.306280] env[63202]: DEBUG oslo_concurrency.lockutils [req-9719aba8-582e-414c-b14e-dfe1481eb9b7 req-c5127086-6e7a-4690-a6a3-a436c187ae08 service nova] Acquired lock "refresh_cache-3080314c-938e-4c27-bffd-547bdc7e6e38" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.306441] env[63202]: DEBUG nova.network.neutron [req-9719aba8-582e-414c-b14e-dfe1481eb9b7 req-c5127086-6e7a-4690-a6a3-a436c187ae08 service nova] [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] Refreshing network info cache for port 3c2c28dc-66c3-479c-a3a2-2d1ed7c3ae4d {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 767.307621] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-acfa610c-d410-4572-b809-3bd280e71ce3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.316526] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e05c13c4-1a4b-43c9-98c5-944996092420 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.338903] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3080314c-938e-4c27-bffd-547bdc7e6e38 could not be found. [ 767.339126] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 767.339311] env[63202]: INFO nova.compute.manager [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] Took 0.03 seconds to destroy the instance on the hypervisor. [ 767.339537] env[63202]: DEBUG oslo.service.loopingcall [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 767.339729] env[63202]: DEBUG nova.compute.manager [-] [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 767.339815] env[63202]: DEBUG nova.network.neutron [-] [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 767.355719] env[63202]: DEBUG nova.network.neutron [-] [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 767.491573] env[63202]: DEBUG nova.compute.manager [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 767.797297] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0938f7dd-d776-40f1-b0a6-0a58ec6d76b6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.804513] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb268c3f-7ac4-4952-b4b1-81a4a65d8463 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.836652] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d839a2a7-13b0-47ad-9b22-ef25415300d5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.843988] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f21eb3f1-7c03-4fb2-88a5-5d0c29fce06c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.857337] env[63202]: DEBUG nova.network.neutron [-] [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.858674] env[63202]: DEBUG nova.compute.provider_tree [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 767.860457] env[63202]: DEBUG nova.network.neutron [req-9719aba8-582e-414c-b14e-dfe1481eb9b7 req-c5127086-6e7a-4690-a6a3-a436c187ae08 service nova] [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 767.914322] env[63202]: DEBUG nova.network.neutron [req-9719aba8-582e-414c-b14e-dfe1481eb9b7 req-c5127086-6e7a-4690-a6a3-a436c187ae08 service nova] [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.999995] env[63202]: INFO nova.virt.block_device [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Booting with volume 9f89c70b-57ad-42ed-8883-bef33fe09c62 at /dev/sda [ 768.046388] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-62544fdb-099d-447a-a288-69054f108fbf {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.055837] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3737a8c5-0175-4f5a-a64d-2a06ab7202d4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.077688] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6a62130b-64f0-447a-a15d-33ed16e1ed3c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.085254] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7981a293-5a12-4581-a2de-ec1aeee38e06 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.106298] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daa65cab-931a-40d7-a7e1-bbd2ddbc714c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.112244] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2661650-a78a-4f3c-9614-ce359c720e7b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.127152] env[63202]: DEBUG nova.virt.block_device [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Updating existing volume attachment record: d08f9dfd-69b8-489a-b4ed-cdc0387232c0 {{(pid=63202) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 768.360619] env[63202]: INFO nova.compute.manager [-] [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] Took 1.02 seconds to deallocate network for instance. [ 768.363203] env[63202]: DEBUG nova.compute.claims [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 768.363382] env[63202]: DEBUG oslo_concurrency.lockutils [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.364149] env[63202]: DEBUG nova.scheduler.client.report [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 768.417563] env[63202]: DEBUG oslo_concurrency.lockutils [req-9719aba8-582e-414c-b14e-dfe1481eb9b7 req-c5127086-6e7a-4690-a6a3-a436c187ae08 service nova] Releasing lock "refresh_cache-3080314c-938e-4c27-bffd-547bdc7e6e38" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.417809] env[63202]: DEBUG nova.compute.manager [req-9719aba8-582e-414c-b14e-dfe1481eb9b7 req-c5127086-6e7a-4690-a6a3-a436c187ae08 service nova] [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] Received event network-vif-deleted-3c2c28dc-66c3-479c-a3a2-2d1ed7c3ae4d {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 768.449443] env[63202]: ERROR nova.compute.manager [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 52847cc7-89c1-4996-92bd-ec6cefb47d70, please check neutron logs for more information. [ 768.449443] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 768.449443] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 768.449443] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 768.449443] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 768.449443] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 768.449443] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 768.449443] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 768.449443] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 768.449443] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 768.449443] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 768.449443] env[63202]: ERROR nova.compute.manager raise self.value [ 768.449443] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 768.449443] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 768.449443] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 768.449443] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 768.450374] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 768.450374] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 768.450374] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 52847cc7-89c1-4996-92bd-ec6cefb47d70, please check neutron logs for more information. [ 768.450374] env[63202]: ERROR nova.compute.manager [ 768.450374] env[63202]: Traceback (most recent call last): [ 768.450374] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 768.450374] env[63202]: listener.cb(fileno) [ 768.450374] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 768.450374] env[63202]: result = function(*args, **kwargs) [ 768.450374] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 768.450374] env[63202]: return func(*args, **kwargs) [ 768.450374] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 768.450374] env[63202]: raise e [ 768.450374] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 768.450374] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 768.450374] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 768.450374] env[63202]: created_port_ids = self._update_ports_for_instance( [ 768.450374] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 768.450374] env[63202]: with excutils.save_and_reraise_exception(): [ 768.450374] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 768.450374] env[63202]: self.force_reraise() [ 768.450374] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 768.450374] env[63202]: raise self.value [ 768.450374] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 768.450374] env[63202]: updated_port = self._update_port( [ 768.450374] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 768.450374] env[63202]: _ensure_no_port_binding_failure(port) [ 768.450374] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 768.450374] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 768.451650] env[63202]: nova.exception.PortBindingFailed: Binding failed for port 52847cc7-89c1-4996-92bd-ec6cefb47d70, please check neutron logs for more information. [ 768.451650] env[63202]: Removing descriptor: 16 [ 768.869315] env[63202]: DEBUG oslo_concurrency.lockutils [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.389s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.869804] env[63202]: DEBUG nova.compute.manager [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: fea91da2-186a-44b6-8e20-535e3b123890] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 768.872326] env[63202]: DEBUG oslo_concurrency.lockutils [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.852s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.873831] env[63202]: INFO nova.compute.claims [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 769.028450] env[63202]: DEBUG nova.compute.manager [req-f0f8e3d6-84fb-4b1a-b001-b15a461cdb07 req-f5ae54fc-be47-48a3-8fc6-b283e3ed967b service nova] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Received event network-changed-52847cc7-89c1-4996-92bd-ec6cefb47d70 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 769.028614] env[63202]: DEBUG nova.compute.manager [req-f0f8e3d6-84fb-4b1a-b001-b15a461cdb07 req-f5ae54fc-be47-48a3-8fc6-b283e3ed967b service nova] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Refreshing instance network info cache due to event network-changed-52847cc7-89c1-4996-92bd-ec6cefb47d70. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 769.028888] env[63202]: DEBUG oslo_concurrency.lockutils [req-f0f8e3d6-84fb-4b1a-b001-b15a461cdb07 req-f5ae54fc-be47-48a3-8fc6-b283e3ed967b service nova] Acquiring lock "refresh_cache-7efc454e-337d-43db-9076-bfc5b89eeea4" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.029040] env[63202]: DEBUG oslo_concurrency.lockutils [req-f0f8e3d6-84fb-4b1a-b001-b15a461cdb07 req-f5ae54fc-be47-48a3-8fc6-b283e3ed967b service nova] Acquired lock "refresh_cache-7efc454e-337d-43db-9076-bfc5b89eeea4" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.029203] env[63202]: DEBUG nova.network.neutron [req-f0f8e3d6-84fb-4b1a-b001-b15a461cdb07 req-f5ae54fc-be47-48a3-8fc6-b283e3ed967b service nova] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Refreshing network info cache for port 52847cc7-89c1-4996-92bd-ec6cefb47d70 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 769.378592] env[63202]: DEBUG nova.compute.utils [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 769.381643] env[63202]: DEBUG nova.compute.manager [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: fea91da2-186a-44b6-8e20-535e3b123890] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 769.381794] env[63202]: DEBUG nova.network.neutron [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: fea91da2-186a-44b6-8e20-535e3b123890] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 769.443242] env[63202]: DEBUG nova.policy [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ec07eaa606564f499b1e735409881c8d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e2926b8bd92644deb614a386f00be6c8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 769.549381] env[63202]: DEBUG nova.network.neutron [req-f0f8e3d6-84fb-4b1a-b001-b15a461cdb07 req-f5ae54fc-be47-48a3-8fc6-b283e3ed967b service nova] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 769.637670] env[63202]: DEBUG nova.network.neutron [req-f0f8e3d6-84fb-4b1a-b001-b15a461cdb07 req-f5ae54fc-be47-48a3-8fc6-b283e3ed967b service nova] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.883090] env[63202]: DEBUG nova.compute.manager [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: fea91da2-186a-44b6-8e20-535e3b123890] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 770.064033] env[63202]: DEBUG nova.network.neutron [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: fea91da2-186a-44b6-8e20-535e3b123890] Successfully created port: 58102740-cab3-4fc7-8dbf-725c4f0131f6 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 770.142267] env[63202]: DEBUG oslo_concurrency.lockutils [req-f0f8e3d6-84fb-4b1a-b001-b15a461cdb07 req-f5ae54fc-be47-48a3-8fc6-b283e3ed967b service nova] Releasing lock "refresh_cache-7efc454e-337d-43db-9076-bfc5b89eeea4" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.142419] env[63202]: DEBUG nova.compute.manager [req-f0f8e3d6-84fb-4b1a-b001-b15a461cdb07 req-f5ae54fc-be47-48a3-8fc6-b283e3ed967b service nova] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Received event network-vif-deleted-52847cc7-89c1-4996-92bd-ec6cefb47d70 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 770.227892] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06a8c391-179e-40cc-aad9-f50f70fb36da {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.232957] env[63202]: DEBUG nova.compute.manager [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 770.234126] env[63202]: DEBUG nova.virt.hardware [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 770.234493] env[63202]: DEBUG nova.virt.hardware [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 770.234564] env[63202]: DEBUG nova.virt.hardware [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 770.234765] env[63202]: DEBUG nova.virt.hardware [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 770.234943] env[63202]: DEBUG nova.virt.hardware [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 770.235179] env[63202]: DEBUG nova.virt.hardware [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 770.235449] env[63202]: DEBUG nova.virt.hardware [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 770.235647] env[63202]: DEBUG nova.virt.hardware [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 770.236612] env[63202]: DEBUG nova.virt.hardware [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 770.236612] env[63202]: DEBUG nova.virt.hardware [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 770.236612] env[63202]: DEBUG nova.virt.hardware [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 770.238780] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1e77c2d-e743-498b-89e8-3ec910fdf627 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.242556] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3cc3dd6-9b2c-4bfe-9733-4067a5df8a88 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.275143] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1deb0ad-7050-4c8e-89e1-f4257241b3d0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.279568] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f3be966-4bc9-41e9-a76c-486a1f9ab500 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.293654] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc369812-5fd3-4de8-8f79-6ba5d729bc6b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.297788] env[63202]: ERROR nova.compute.manager [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 52847cc7-89c1-4996-92bd-ec6cefb47d70, please check neutron logs for more information. [ 770.297788] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Traceback (most recent call last): [ 770.297788] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 770.297788] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] yield resources [ 770.297788] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 770.297788] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] self.driver.spawn(context, instance, image_meta, [ 770.297788] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 770.297788] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 770.297788] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 770.297788] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] vm_ref = self.build_virtual_machine(instance, [ 770.297788] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 770.298238] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] vif_infos = vmwarevif.get_vif_info(self._session, [ 770.298238] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 770.298238] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] for vif in network_info: [ 770.298238] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 770.298238] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] return self._sync_wrapper(fn, *args, **kwargs) [ 770.298238] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 770.298238] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] self.wait() [ 770.298238] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 770.298238] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] self[:] = self._gt.wait() [ 770.298238] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 770.298238] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] return self._exit_event.wait() [ 770.298238] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 770.298238] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] current.throw(*self._exc) [ 770.298687] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 770.298687] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] result = function(*args, **kwargs) [ 770.298687] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 770.298687] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] return func(*args, **kwargs) [ 770.298687] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 770.298687] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] raise e [ 770.298687] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 770.298687] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] nwinfo = self.network_api.allocate_for_instance( [ 770.298687] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 770.298687] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] created_port_ids = self._update_ports_for_instance( [ 770.298687] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 770.298687] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] with excutils.save_and_reraise_exception(): [ 770.298687] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 770.299191] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] self.force_reraise() [ 770.299191] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 770.299191] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] raise self.value [ 770.299191] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 770.299191] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] updated_port = self._update_port( [ 770.299191] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 770.299191] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] _ensure_no_port_binding_failure(port) [ 770.299191] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 770.299191] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] raise exception.PortBindingFailed(port_id=port['id']) [ 770.299191] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] nova.exception.PortBindingFailed: Binding failed for port 52847cc7-89c1-4996-92bd-ec6cefb47d70, please check neutron logs for more information. [ 770.299191] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] [ 770.299191] env[63202]: INFO nova.compute.manager [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Terminating instance [ 770.300985] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] Acquiring lock "refresh_cache-7efc454e-337d-43db-9076-bfc5b89eeea4" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 770.300985] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] Acquired lock "refresh_cache-7efc454e-337d-43db-9076-bfc5b89eeea4" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.300985] env[63202]: DEBUG nova.network.neutron [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 770.310040] env[63202]: DEBUG nova.compute.provider_tree [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 770.814568] env[63202]: DEBUG nova.scheduler.client.report [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 770.836977] env[63202]: DEBUG nova.network.neutron [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 770.896414] env[63202]: DEBUG nova.compute.manager [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: fea91da2-186a-44b6-8e20-535e3b123890] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 770.925938] env[63202]: DEBUG nova.network.neutron [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.929091] env[63202]: DEBUG nova.virt.hardware [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 770.929217] env[63202]: DEBUG nova.virt.hardware [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 770.929321] env[63202]: DEBUG nova.virt.hardware [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 770.929502] env[63202]: DEBUG nova.virt.hardware [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 770.929643] env[63202]: DEBUG nova.virt.hardware [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 770.929784] env[63202]: DEBUG nova.virt.hardware [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 770.930036] env[63202]: DEBUG nova.virt.hardware [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 770.930189] env[63202]: DEBUG nova.virt.hardware [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 770.930351] env[63202]: DEBUG nova.virt.hardware [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 770.930505] env[63202]: DEBUG nova.virt.hardware [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 770.930737] env[63202]: DEBUG nova.virt.hardware [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 770.931613] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f517621d-98cb-4536-89be-c1594aa3297a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.941017] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04f47a7f-39bd-46bd-8204-9217ec07fd6f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.090329] env[63202]: DEBUG nova.compute.manager [req-aa7f1c08-852d-47d1-94a2-88334c16f680 req-cfba1416-c208-495d-bfbb-794cbf6beb63 service nova] [instance: fea91da2-186a-44b6-8e20-535e3b123890] Received event network-changed-58102740-cab3-4fc7-8dbf-725c4f0131f6 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 771.090819] env[63202]: DEBUG nova.compute.manager [req-aa7f1c08-852d-47d1-94a2-88334c16f680 req-cfba1416-c208-495d-bfbb-794cbf6beb63 service nova] [instance: fea91da2-186a-44b6-8e20-535e3b123890] Refreshing instance network info cache due to event network-changed-58102740-cab3-4fc7-8dbf-725c4f0131f6. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 771.091208] env[63202]: DEBUG oslo_concurrency.lockutils [req-aa7f1c08-852d-47d1-94a2-88334c16f680 req-cfba1416-c208-495d-bfbb-794cbf6beb63 service nova] Acquiring lock "refresh_cache-fea91da2-186a-44b6-8e20-535e3b123890" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 771.091348] env[63202]: DEBUG oslo_concurrency.lockutils [req-aa7f1c08-852d-47d1-94a2-88334c16f680 req-cfba1416-c208-495d-bfbb-794cbf6beb63 service nova] Acquired lock "refresh_cache-fea91da2-186a-44b6-8e20-535e3b123890" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.091565] env[63202]: DEBUG nova.network.neutron [req-aa7f1c08-852d-47d1-94a2-88334c16f680 req-cfba1416-c208-495d-bfbb-794cbf6beb63 service nova] [instance: fea91da2-186a-44b6-8e20-535e3b123890] Refreshing network info cache for port 58102740-cab3-4fc7-8dbf-725c4f0131f6 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 771.101256] env[63202]: ERROR nova.compute.manager [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 58102740-cab3-4fc7-8dbf-725c4f0131f6, please check neutron logs for more information. [ 771.101256] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 771.101256] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 771.101256] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 771.101256] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 771.101256] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 771.101256] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 771.101256] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 771.101256] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 771.101256] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 771.101256] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 771.101256] env[63202]: ERROR nova.compute.manager raise self.value [ 771.101256] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 771.101256] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 771.101256] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 771.101256] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 771.101725] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 771.101725] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 771.101725] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 58102740-cab3-4fc7-8dbf-725c4f0131f6, please check neutron logs for more information. [ 771.101725] env[63202]: ERROR nova.compute.manager [ 771.101725] env[63202]: Traceback (most recent call last): [ 771.101725] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 771.101725] env[63202]: listener.cb(fileno) [ 771.101725] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 771.101725] env[63202]: result = function(*args, **kwargs) [ 771.101725] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 771.101725] env[63202]: return func(*args, **kwargs) [ 771.101725] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 771.101725] env[63202]: raise e [ 771.101725] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 771.101725] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 771.101725] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 771.101725] env[63202]: created_port_ids = self._update_ports_for_instance( [ 771.101725] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 771.101725] env[63202]: with excutils.save_and_reraise_exception(): [ 771.101725] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 771.101725] env[63202]: self.force_reraise() [ 771.101725] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 771.101725] env[63202]: raise self.value [ 771.101725] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 771.101725] env[63202]: updated_port = self._update_port( [ 771.101725] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 771.101725] env[63202]: _ensure_no_port_binding_failure(port) [ 771.101725] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 771.101725] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 771.102836] env[63202]: nova.exception.PortBindingFailed: Binding failed for port 58102740-cab3-4fc7-8dbf-725c4f0131f6, please check neutron logs for more information. [ 771.102836] env[63202]: Removing descriptor: 16 [ 771.102836] env[63202]: ERROR nova.compute.manager [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: fea91da2-186a-44b6-8e20-535e3b123890] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 58102740-cab3-4fc7-8dbf-725c4f0131f6, please check neutron logs for more information. [ 771.102836] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] Traceback (most recent call last): [ 771.102836] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 771.102836] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] yield resources [ 771.102836] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 771.102836] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] self.driver.spawn(context, instance, image_meta, [ 771.102836] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 771.102836] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] self._vmops.spawn(context, instance, image_meta, injected_files, [ 771.102836] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 771.102836] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] vm_ref = self.build_virtual_machine(instance, [ 771.103243] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 771.103243] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] vif_infos = vmwarevif.get_vif_info(self._session, [ 771.103243] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 771.103243] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] for vif in network_info: [ 771.103243] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 771.103243] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] return self._sync_wrapper(fn, *args, **kwargs) [ 771.103243] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 771.103243] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] self.wait() [ 771.103243] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 771.103243] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] self[:] = self._gt.wait() [ 771.103243] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 771.103243] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] return self._exit_event.wait() [ 771.103243] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 771.103654] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] result = hub.switch() [ 771.103654] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 771.103654] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] return self.greenlet.switch() [ 771.103654] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 771.103654] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] result = function(*args, **kwargs) [ 771.103654] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 771.103654] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] return func(*args, **kwargs) [ 771.103654] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 771.103654] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] raise e [ 771.103654] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 771.103654] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] nwinfo = self.network_api.allocate_for_instance( [ 771.103654] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 771.103654] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] created_port_ids = self._update_ports_for_instance( [ 771.104073] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 771.104073] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] with excutils.save_and_reraise_exception(): [ 771.104073] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 771.104073] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] self.force_reraise() [ 771.104073] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 771.104073] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] raise self.value [ 771.104073] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 771.104073] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] updated_port = self._update_port( [ 771.104073] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 771.104073] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] _ensure_no_port_binding_failure(port) [ 771.104073] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 771.104073] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] raise exception.PortBindingFailed(port_id=port['id']) [ 771.104448] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] nova.exception.PortBindingFailed: Binding failed for port 58102740-cab3-4fc7-8dbf-725c4f0131f6, please check neutron logs for more information. [ 771.104448] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] [ 771.104448] env[63202]: INFO nova.compute.manager [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: fea91da2-186a-44b6-8e20-535e3b123890] Terminating instance [ 771.104849] env[63202]: DEBUG oslo_concurrency.lockutils [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Acquiring lock "refresh_cache-fea91da2-186a-44b6-8e20-535e3b123890" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 771.321447] env[63202]: DEBUG oslo_concurrency.lockutils [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.449s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.321997] env[63202]: DEBUG nova.compute.manager [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 771.324579] env[63202]: DEBUG oslo_concurrency.lockutils [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.362s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.326011] env[63202]: INFO nova.compute.claims [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 771.436411] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] Releasing lock "refresh_cache-7efc454e-337d-43db-9076-bfc5b89eeea4" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.437061] env[63202]: DEBUG nova.compute.manager [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 771.437384] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1ec90242-17e7-490b-8e7b-51582fe5eb50 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.447342] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cacf8f38-a425-4c08-9b2c-6179a76b99f3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.468851] env[63202]: WARNING nova.virt.vmwareapi.driver [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 7efc454e-337d-43db-9076-bfc5b89eeea4 could not be found. [ 771.469054] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 771.469339] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5994e7f4-ec1c-4fb4-83ec-bd3da3c88c51 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.476795] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd0c6fcf-4598-4140-a376-3762ebd085e4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.497489] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7efc454e-337d-43db-9076-bfc5b89eeea4 could not be found. [ 771.497699] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 771.497874] env[63202]: INFO nova.compute.manager [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Took 0.06 seconds to destroy the instance on the hypervisor. [ 771.498120] env[63202]: DEBUG oslo.service.loopingcall [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 771.498333] env[63202]: DEBUG nova.compute.manager [-] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 771.498424] env[63202]: DEBUG nova.network.neutron [-] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 771.513166] env[63202]: DEBUG nova.network.neutron [-] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 771.610391] env[63202]: DEBUG nova.network.neutron [req-aa7f1c08-852d-47d1-94a2-88334c16f680 req-cfba1416-c208-495d-bfbb-794cbf6beb63 service nova] [instance: fea91da2-186a-44b6-8e20-535e3b123890] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 771.667581] env[63202]: DEBUG nova.network.neutron [req-aa7f1c08-852d-47d1-94a2-88334c16f680 req-cfba1416-c208-495d-bfbb-794cbf6beb63 service nova] [instance: fea91da2-186a-44b6-8e20-535e3b123890] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 771.832853] env[63202]: DEBUG nova.compute.utils [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 771.834435] env[63202]: DEBUG nova.compute.manager [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 771.834435] env[63202]: DEBUG nova.network.neutron [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 771.872411] env[63202]: DEBUG nova.policy [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2bd91973e1bb4ce7be8bce9bfe34a941', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e972faeaedc6468aab7e7cfee88a477b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 772.015805] env[63202]: DEBUG nova.network.neutron [-] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.171155] env[63202]: DEBUG oslo_concurrency.lockutils [req-aa7f1c08-852d-47d1-94a2-88334c16f680 req-cfba1416-c208-495d-bfbb-794cbf6beb63 service nova] Releasing lock "refresh_cache-fea91da2-186a-44b6-8e20-535e3b123890" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 772.171569] env[63202]: DEBUG oslo_concurrency.lockutils [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Acquired lock "refresh_cache-fea91da2-186a-44b6-8e20-535e3b123890" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.171752] env[63202]: DEBUG nova.network.neutron [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: fea91da2-186a-44b6-8e20-535e3b123890] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 772.210188] env[63202]: DEBUG nova.network.neutron [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] Successfully created port: 0ab92838-e88e-4996-94b8-21a3a09753d8 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 772.340368] env[63202]: DEBUG nova.compute.manager [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 772.518105] env[63202]: INFO nova.compute.manager [-] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Took 1.02 seconds to deallocate network for instance. [ 772.621048] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80dba1c3-e0cb-440e-b696-cd5e7df06e2c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.627961] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad155f4f-7a53-48c7-bd5d-04c1294c85ba {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.657310] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78a38ab1-fc4e-431c-8456-97ac20f74ea9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.663949] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21753e25-0cba-4ff3-be02-ea517696a941 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.678644] env[63202]: DEBUG nova.compute.provider_tree [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 772.693354] env[63202]: DEBUG nova.network.neutron [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: fea91da2-186a-44b6-8e20-535e3b123890] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 772.740858] env[63202]: DEBUG nova.network.neutron [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: fea91da2-186a-44b6-8e20-535e3b123890] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.089837] env[63202]: INFO nova.compute.manager [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Took 0.57 seconds to detach 1 volumes for instance. [ 773.091979] env[63202]: DEBUG nova.compute.claims [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 773.092174] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 773.175433] env[63202]: DEBUG nova.compute.manager [req-91d0190b-0b23-4c46-a55c-ab6759803a67 req-1a198e72-0e29-46fa-bf30-80ed888b0c00 service nova] [instance: fea91da2-186a-44b6-8e20-535e3b123890] Received event network-vif-deleted-58102740-cab3-4fc7-8dbf-725c4f0131f6 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 773.175433] env[63202]: DEBUG nova.compute.manager [req-91d0190b-0b23-4c46-a55c-ab6759803a67 req-1a198e72-0e29-46fa-bf30-80ed888b0c00 service nova] [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] Received event network-changed-0ab92838-e88e-4996-94b8-21a3a09753d8 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 773.175433] env[63202]: DEBUG nova.compute.manager [req-91d0190b-0b23-4c46-a55c-ab6759803a67 req-1a198e72-0e29-46fa-bf30-80ed888b0c00 service nova] [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] Refreshing instance network info cache due to event network-changed-0ab92838-e88e-4996-94b8-21a3a09753d8. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 773.175433] env[63202]: DEBUG oslo_concurrency.lockutils [req-91d0190b-0b23-4c46-a55c-ab6759803a67 req-1a198e72-0e29-46fa-bf30-80ed888b0c00 service nova] Acquiring lock "refresh_cache-8045881e-9bce-46e7-98c6-a7989f61a31e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.175433] env[63202]: DEBUG oslo_concurrency.lockutils [req-91d0190b-0b23-4c46-a55c-ab6759803a67 req-1a198e72-0e29-46fa-bf30-80ed888b0c00 service nova] Acquired lock "refresh_cache-8045881e-9bce-46e7-98c6-a7989f61a31e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.175798] env[63202]: DEBUG nova.network.neutron [req-91d0190b-0b23-4c46-a55c-ab6759803a67 req-1a198e72-0e29-46fa-bf30-80ed888b0c00 service nova] [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] Refreshing network info cache for port 0ab92838-e88e-4996-94b8-21a3a09753d8 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 773.181408] env[63202]: DEBUG nova.scheduler.client.report [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 773.215498] env[63202]: ERROR nova.compute.manager [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0ab92838-e88e-4996-94b8-21a3a09753d8, please check neutron logs for more information. [ 773.215498] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 773.215498] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 773.215498] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 773.215498] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 773.215498] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 773.215498] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 773.215498] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 773.215498] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 773.215498] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 773.215498] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 773.215498] env[63202]: ERROR nova.compute.manager raise self.value [ 773.215498] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 773.215498] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 773.215498] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 773.215498] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 773.215955] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 773.215955] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 773.215955] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0ab92838-e88e-4996-94b8-21a3a09753d8, please check neutron logs for more information. [ 773.215955] env[63202]: ERROR nova.compute.manager [ 773.215955] env[63202]: Traceback (most recent call last): [ 773.215955] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 773.215955] env[63202]: listener.cb(fileno) [ 773.215955] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 773.215955] env[63202]: result = function(*args, **kwargs) [ 773.215955] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 773.215955] env[63202]: return func(*args, **kwargs) [ 773.215955] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 773.215955] env[63202]: raise e [ 773.215955] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 773.215955] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 773.215955] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 773.215955] env[63202]: created_port_ids = self._update_ports_for_instance( [ 773.215955] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 773.215955] env[63202]: with excutils.save_and_reraise_exception(): [ 773.215955] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 773.215955] env[63202]: self.force_reraise() [ 773.215955] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 773.215955] env[63202]: raise self.value [ 773.215955] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 773.215955] env[63202]: updated_port = self._update_port( [ 773.215955] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 773.215955] env[63202]: _ensure_no_port_binding_failure(port) [ 773.215955] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 773.215955] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 773.216712] env[63202]: nova.exception.PortBindingFailed: Binding failed for port 0ab92838-e88e-4996-94b8-21a3a09753d8, please check neutron logs for more information. [ 773.216712] env[63202]: Removing descriptor: 16 [ 773.246290] env[63202]: DEBUG oslo_concurrency.lockutils [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Releasing lock "refresh_cache-fea91da2-186a-44b6-8e20-535e3b123890" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.246689] env[63202]: DEBUG nova.compute.manager [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: fea91da2-186a-44b6-8e20-535e3b123890] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 773.246879] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: fea91da2-186a-44b6-8e20-535e3b123890] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 773.247181] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c3424f49-98a2-4708-9046-560f8f3aef17 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.257406] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1af000c-6ed2-4587-922a-790134e2ffec {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.279442] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: fea91da2-186a-44b6-8e20-535e3b123890] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance fea91da2-186a-44b6-8e20-535e3b123890 could not be found. [ 773.279659] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: fea91da2-186a-44b6-8e20-535e3b123890] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 773.279837] env[63202]: INFO nova.compute.manager [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: fea91da2-186a-44b6-8e20-535e3b123890] Took 0.03 seconds to destroy the instance on the hypervisor. [ 773.280088] env[63202]: DEBUG oslo.service.loopingcall [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 773.280291] env[63202]: DEBUG nova.compute.manager [-] [instance: fea91da2-186a-44b6-8e20-535e3b123890] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 773.280383] env[63202]: DEBUG nova.network.neutron [-] [instance: fea91da2-186a-44b6-8e20-535e3b123890] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 773.292837] env[63202]: DEBUG nova.network.neutron [-] [instance: fea91da2-186a-44b6-8e20-535e3b123890] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 773.351648] env[63202]: DEBUG nova.compute.manager [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 773.375918] env[63202]: DEBUG nova.virt.hardware [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 773.376208] env[63202]: DEBUG nova.virt.hardware [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 773.376408] env[63202]: DEBUG nova.virt.hardware [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 773.376629] env[63202]: DEBUG nova.virt.hardware [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 773.376797] env[63202]: DEBUG nova.virt.hardware [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 773.376946] env[63202]: DEBUG nova.virt.hardware [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 773.377169] env[63202]: DEBUG nova.virt.hardware [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 773.377317] env[63202]: DEBUG nova.virt.hardware [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 773.377473] env[63202]: DEBUG nova.virt.hardware [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 773.377649] env[63202]: DEBUG nova.virt.hardware [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 773.377796] env[63202]: DEBUG nova.virt.hardware [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 773.378666] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1c8fd9d-a2b8-4886-beb0-fa5e7ef4b0a2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.386396] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65103dd2-8dbd-4644-93d6-9a537d6fb253 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.400108] env[63202]: ERROR nova.compute.manager [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0ab92838-e88e-4996-94b8-21a3a09753d8, please check neutron logs for more information. [ 773.400108] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] Traceback (most recent call last): [ 773.400108] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 773.400108] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] yield resources [ 773.400108] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 773.400108] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] self.driver.spawn(context, instance, image_meta, [ 773.400108] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 773.400108] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 773.400108] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 773.400108] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] vm_ref = self.build_virtual_machine(instance, [ 773.400108] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 773.400523] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] vif_infos = vmwarevif.get_vif_info(self._session, [ 773.400523] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 773.400523] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] for vif in network_info: [ 773.400523] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 773.400523] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] return self._sync_wrapper(fn, *args, **kwargs) [ 773.400523] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 773.400523] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] self.wait() [ 773.400523] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 773.400523] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] self[:] = self._gt.wait() [ 773.400523] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 773.400523] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] return self._exit_event.wait() [ 773.400523] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 773.400523] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] current.throw(*self._exc) [ 773.400931] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 773.400931] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] result = function(*args, **kwargs) [ 773.400931] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 773.400931] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] return func(*args, **kwargs) [ 773.400931] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 773.400931] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] raise e [ 773.400931] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 773.400931] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] nwinfo = self.network_api.allocate_for_instance( [ 773.400931] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 773.400931] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] created_port_ids = self._update_ports_for_instance( [ 773.400931] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 773.400931] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] with excutils.save_and_reraise_exception(): [ 773.400931] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 773.401669] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] self.force_reraise() [ 773.401669] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 773.401669] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] raise self.value [ 773.401669] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 773.401669] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] updated_port = self._update_port( [ 773.401669] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 773.401669] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] _ensure_no_port_binding_failure(port) [ 773.401669] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 773.401669] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] raise exception.PortBindingFailed(port_id=port['id']) [ 773.401669] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] nova.exception.PortBindingFailed: Binding failed for port 0ab92838-e88e-4996-94b8-21a3a09753d8, please check neutron logs for more information. [ 773.401669] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] [ 773.401669] env[63202]: INFO nova.compute.manager [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] Terminating instance [ 773.402384] env[63202]: DEBUG oslo_concurrency.lockutils [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "refresh_cache-8045881e-9bce-46e7-98c6-a7989f61a31e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.686669] env[63202]: DEBUG oslo_concurrency.lockutils [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.362s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.687170] env[63202]: DEBUG nova.compute.manager [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 773.689705] env[63202]: DEBUG oslo_concurrency.lockutils [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.877s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 773.691934] env[63202]: INFO nova.compute.claims [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 773.694848] env[63202]: DEBUG nova.network.neutron [req-91d0190b-0b23-4c46-a55c-ab6759803a67 req-1a198e72-0e29-46fa-bf30-80ed888b0c00 service nova] [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 773.772114] env[63202]: DEBUG nova.network.neutron [req-91d0190b-0b23-4c46-a55c-ab6759803a67 req-1a198e72-0e29-46fa-bf30-80ed888b0c00 service nova] [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.795075] env[63202]: DEBUG nova.network.neutron [-] [instance: fea91da2-186a-44b6-8e20-535e3b123890] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.199114] env[63202]: DEBUG nova.compute.utils [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 774.203238] env[63202]: DEBUG nova.compute.manager [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 774.203400] env[63202]: DEBUG nova.network.neutron [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 774.259517] env[63202]: DEBUG nova.policy [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '134555416ca4458c9f144986058ad1c4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a4e9d5e5bd45495ea03330115395aa5d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 774.274753] env[63202]: DEBUG oslo_concurrency.lockutils [req-91d0190b-0b23-4c46-a55c-ab6759803a67 req-1a198e72-0e29-46fa-bf30-80ed888b0c00 service nova] Releasing lock "refresh_cache-8045881e-9bce-46e7-98c6-a7989f61a31e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 774.275355] env[63202]: DEBUG oslo_concurrency.lockutils [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquired lock "refresh_cache-8045881e-9bce-46e7-98c6-a7989f61a31e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 774.275355] env[63202]: DEBUG nova.network.neutron [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 774.297291] env[63202]: INFO nova.compute.manager [-] [instance: fea91da2-186a-44b6-8e20-535e3b123890] Took 1.02 seconds to deallocate network for instance. [ 774.299420] env[63202]: DEBUG nova.compute.claims [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: fea91da2-186a-44b6-8e20-535e3b123890] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 774.299564] env[63202]: DEBUG oslo_concurrency.lockutils [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.677091] env[63202]: DEBUG nova.network.neutron [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] Successfully created port: ad906a1e-35dc-4bb7-85b6-2672972103d0 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 774.707016] env[63202]: DEBUG nova.compute.manager [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 774.800151] env[63202]: DEBUG nova.network.neutron [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 774.897311] env[63202]: DEBUG nova.network.neutron [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.074838] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69f47fb1-63d2-4574-8889-96f03ab13f24 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.085725] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b2f9cc2-fe4c-40be-8ea0-619479270969 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.135745] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-876869e6-de6a-4e97-944a-245a618805d4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.147551] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5edf2e4c-b363-49f3-9f66-48d07a4573da {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.170198] env[63202]: DEBUG nova.compute.provider_tree [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 775.201987] env[63202]: DEBUG nova.compute.manager [req-79c6fc2d-8ac0-4f88-a95a-df47cacea5ea req-91e0fc31-ecad-4c40-8af6-223107a7384d service nova] [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] Received event network-vif-deleted-0ab92838-e88e-4996-94b8-21a3a09753d8 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 775.402640] env[63202]: DEBUG oslo_concurrency.lockutils [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Releasing lock "refresh_cache-8045881e-9bce-46e7-98c6-a7989f61a31e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.403069] env[63202]: DEBUG nova.compute.manager [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 775.403266] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 775.403557] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-422a1531-db4b-48ac-a815-13723ba1eea0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.415022] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aa477a4-57c6-4ad0-af63-8a32aa4113f5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.441773] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8045881e-9bce-46e7-98c6-a7989f61a31e could not be found. [ 775.442010] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 775.442209] env[63202]: INFO nova.compute.manager [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 775.442460] env[63202]: DEBUG oslo.service.loopingcall [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 775.442688] env[63202]: DEBUG nova.compute.manager [-] [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 775.442782] env[63202]: DEBUG nova.network.neutron [-] [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 775.461189] env[63202]: DEBUG nova.network.neutron [-] [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 775.674491] env[63202]: DEBUG nova.scheduler.client.report [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 775.716675] env[63202]: DEBUG nova.compute.manager [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 775.746572] env[63202]: DEBUG nova.virt.hardware [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 775.746696] env[63202]: DEBUG nova.virt.hardware [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 775.746812] env[63202]: DEBUG nova.virt.hardware [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 775.747212] env[63202]: DEBUG nova.virt.hardware [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 775.747375] env[63202]: DEBUG nova.virt.hardware [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 775.747525] env[63202]: DEBUG nova.virt.hardware [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 775.747736] env[63202]: DEBUG nova.virt.hardware [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 775.748639] env[63202]: DEBUG nova.virt.hardware [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 775.748639] env[63202]: DEBUG nova.virt.hardware [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 775.748639] env[63202]: DEBUG nova.virt.hardware [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 775.748639] env[63202]: DEBUG nova.virt.hardware [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 775.749462] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7686ae0b-b760-4df1-8c40-2632f56c2ed1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.757759] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c5ca5e4-75c6-4e8d-be8d-d0d973e2643d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.874267] env[63202]: ERROR nova.compute.manager [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ad906a1e-35dc-4bb7-85b6-2672972103d0, please check neutron logs for more information. [ 775.874267] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 775.874267] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 775.874267] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 775.874267] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 775.874267] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 775.874267] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 775.874267] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 775.874267] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 775.874267] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 775.874267] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 775.874267] env[63202]: ERROR nova.compute.manager raise self.value [ 775.874267] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 775.874267] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 775.874267] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 775.874267] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 775.874795] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 775.874795] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 775.874795] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ad906a1e-35dc-4bb7-85b6-2672972103d0, please check neutron logs for more information. [ 775.874795] env[63202]: ERROR nova.compute.manager [ 775.874795] env[63202]: Traceback (most recent call last): [ 775.874795] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 775.874795] env[63202]: listener.cb(fileno) [ 775.874795] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 775.874795] env[63202]: result = function(*args, **kwargs) [ 775.874795] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 775.874795] env[63202]: return func(*args, **kwargs) [ 775.874795] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 775.874795] env[63202]: raise e [ 775.874795] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 775.874795] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 775.874795] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 775.874795] env[63202]: created_port_ids = self._update_ports_for_instance( [ 775.874795] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 775.874795] env[63202]: with excutils.save_and_reraise_exception(): [ 775.874795] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 775.874795] env[63202]: self.force_reraise() [ 775.874795] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 775.874795] env[63202]: raise self.value [ 775.874795] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 775.874795] env[63202]: updated_port = self._update_port( [ 775.874795] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 775.874795] env[63202]: _ensure_no_port_binding_failure(port) [ 775.874795] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 775.874795] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 775.875803] env[63202]: nova.exception.PortBindingFailed: Binding failed for port ad906a1e-35dc-4bb7-85b6-2672972103d0, please check neutron logs for more information. [ 775.875803] env[63202]: Removing descriptor: 16 [ 775.875803] env[63202]: ERROR nova.compute.manager [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ad906a1e-35dc-4bb7-85b6-2672972103d0, please check neutron logs for more information. [ 775.875803] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] Traceback (most recent call last): [ 775.875803] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 775.875803] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] yield resources [ 775.875803] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 775.875803] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] self.driver.spawn(context, instance, image_meta, [ 775.875803] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 775.875803] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] self._vmops.spawn(context, instance, image_meta, injected_files, [ 775.875803] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 775.875803] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] vm_ref = self.build_virtual_machine(instance, [ 775.876198] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 775.876198] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] vif_infos = vmwarevif.get_vif_info(self._session, [ 775.876198] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 775.876198] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] for vif in network_info: [ 775.876198] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 775.876198] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] return self._sync_wrapper(fn, *args, **kwargs) [ 775.876198] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 775.876198] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] self.wait() [ 775.876198] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 775.876198] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] self[:] = self._gt.wait() [ 775.876198] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 775.876198] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] return self._exit_event.wait() [ 775.876198] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 775.876604] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] result = hub.switch() [ 775.876604] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 775.876604] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] return self.greenlet.switch() [ 775.876604] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 775.876604] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] result = function(*args, **kwargs) [ 775.876604] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 775.876604] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] return func(*args, **kwargs) [ 775.876604] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 775.876604] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] raise e [ 775.876604] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 775.876604] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] nwinfo = self.network_api.allocate_for_instance( [ 775.876604] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 775.876604] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] created_port_ids = self._update_ports_for_instance( [ 775.876967] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 775.876967] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] with excutils.save_and_reraise_exception(): [ 775.876967] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 775.876967] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] self.force_reraise() [ 775.876967] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 775.876967] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] raise self.value [ 775.876967] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 775.876967] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] updated_port = self._update_port( [ 775.876967] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 775.876967] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] _ensure_no_port_binding_failure(port) [ 775.876967] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 775.876967] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] raise exception.PortBindingFailed(port_id=port['id']) [ 775.877359] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] nova.exception.PortBindingFailed: Binding failed for port ad906a1e-35dc-4bb7-85b6-2672972103d0, please check neutron logs for more information. [ 775.877359] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] [ 775.877359] env[63202]: INFO nova.compute.manager [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] Terminating instance [ 775.877461] env[63202]: DEBUG oslo_concurrency.lockutils [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] Acquiring lock "refresh_cache-e3a6ad78-4f46-42d4-935c-3cf310123530" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 775.877582] env[63202]: DEBUG oslo_concurrency.lockutils [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] Acquired lock "refresh_cache-e3a6ad78-4f46-42d4-935c-3cf310123530" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.877730] env[63202]: DEBUG nova.network.neutron [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 775.963342] env[63202]: DEBUG nova.network.neutron [-] [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 776.184901] env[63202]: DEBUG oslo_concurrency.lockutils [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.495s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.185495] env[63202]: DEBUG nova.compute.manager [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 776.187947] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.042s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.408660] env[63202]: DEBUG nova.network.neutron [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 776.465734] env[63202]: INFO nova.compute.manager [-] [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] Took 1.02 seconds to deallocate network for instance. [ 776.468207] env[63202]: DEBUG nova.compute.claims [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 776.468383] env[63202]: DEBUG oslo_concurrency.lockutils [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.479563] env[63202]: DEBUG nova.network.neutron [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 776.692460] env[63202]: DEBUG nova.compute.utils [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 776.693842] env[63202]: DEBUG nova.compute.manager [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 776.694015] env[63202]: DEBUG nova.network.neutron [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 776.736918] env[63202]: DEBUG nova.policy [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2d26ff71671948518d7ae11efd247a4f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1228ad38d4f9419aa0699bbeeba5c0b7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 776.974810] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-313efc96-ceec-4858-857a-4173ea8cd83d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.981321] env[63202]: DEBUG oslo_concurrency.lockutils [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] Releasing lock "refresh_cache-e3a6ad78-4f46-42d4-935c-3cf310123530" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.981651] env[63202]: DEBUG nova.compute.manager [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 776.981846] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 776.982575] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-46d7152f-f691-487e-b822-6c180ab074a1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.991278] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94454093-a5c3-4254-9833-fea305247aae {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.999945] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8caefa5d-35af-43fa-bab5-4781508319fb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.037510] env[63202]: DEBUG nova.network.neutron [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] Successfully created port: f6092551-11d9-42f1-bd5f-f010c6f57650 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 777.040106] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bbb7b75-0ec8-43dc-80b8-193d96383c61 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.047036] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84a53930-23d1-4138-a58f-900791f68ad5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.054782] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e3a6ad78-4f46-42d4-935c-3cf310123530 could not be found. [ 777.054979] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 777.055171] env[63202]: INFO nova.compute.manager [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] Took 0.07 seconds to destroy the instance on the hypervisor. [ 777.055393] env[63202]: DEBUG oslo.service.loopingcall [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 777.055929] env[63202]: DEBUG nova.compute.manager [-] [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 777.056056] env[63202]: DEBUG nova.network.neutron [-] [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 777.065078] env[63202]: DEBUG nova.compute.provider_tree [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 777.079047] env[63202]: DEBUG nova.network.neutron [-] [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 777.200062] env[63202]: DEBUG nova.compute.manager [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 777.260217] env[63202]: DEBUG nova.compute.manager [req-914cf3a3-8195-4ba6-8ca7-097c9226a2eb req-168d95e7-4882-46b5-975b-0e556f6eb3ec service nova] [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] Received event network-changed-ad906a1e-35dc-4bb7-85b6-2672972103d0 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 777.260499] env[63202]: DEBUG nova.compute.manager [req-914cf3a3-8195-4ba6-8ca7-097c9226a2eb req-168d95e7-4882-46b5-975b-0e556f6eb3ec service nova] [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] Refreshing instance network info cache due to event network-changed-ad906a1e-35dc-4bb7-85b6-2672972103d0. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 777.260720] env[63202]: DEBUG oslo_concurrency.lockutils [req-914cf3a3-8195-4ba6-8ca7-097c9226a2eb req-168d95e7-4882-46b5-975b-0e556f6eb3ec service nova] Acquiring lock "refresh_cache-e3a6ad78-4f46-42d4-935c-3cf310123530" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.260859] env[63202]: DEBUG oslo_concurrency.lockutils [req-914cf3a3-8195-4ba6-8ca7-097c9226a2eb req-168d95e7-4882-46b5-975b-0e556f6eb3ec service nova] Acquired lock "refresh_cache-e3a6ad78-4f46-42d4-935c-3cf310123530" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.261024] env[63202]: DEBUG nova.network.neutron [req-914cf3a3-8195-4ba6-8ca7-097c9226a2eb req-168d95e7-4882-46b5-975b-0e556f6eb3ec service nova] [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] Refreshing network info cache for port ad906a1e-35dc-4bb7-85b6-2672972103d0 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 777.568711] env[63202]: DEBUG nova.scheduler.client.report [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 777.581724] env[63202]: DEBUG nova.network.neutron [-] [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.780667] env[63202]: DEBUG nova.network.neutron [req-914cf3a3-8195-4ba6-8ca7-097c9226a2eb req-168d95e7-4882-46b5-975b-0e556f6eb3ec service nova] [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 777.848014] env[63202]: DEBUG nova.network.neutron [req-914cf3a3-8195-4ba6-8ca7-097c9226a2eb req-168d95e7-4882-46b5-975b-0e556f6eb3ec service nova] [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.057442] env[63202]: ERROR nova.compute.manager [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f6092551-11d9-42f1-bd5f-f010c6f57650, please check neutron logs for more information. [ 778.057442] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 778.057442] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 778.057442] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 778.057442] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 778.057442] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 778.057442] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 778.057442] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 778.057442] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 778.057442] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 778.057442] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 778.057442] env[63202]: ERROR nova.compute.manager raise self.value [ 778.057442] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 778.057442] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 778.057442] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 778.057442] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 778.058351] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 778.058351] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 778.058351] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f6092551-11d9-42f1-bd5f-f010c6f57650, please check neutron logs for more information. [ 778.058351] env[63202]: ERROR nova.compute.manager [ 778.058351] env[63202]: Traceback (most recent call last): [ 778.058351] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 778.058351] env[63202]: listener.cb(fileno) [ 778.058351] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 778.058351] env[63202]: result = function(*args, **kwargs) [ 778.058351] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 778.058351] env[63202]: return func(*args, **kwargs) [ 778.058351] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 778.058351] env[63202]: raise e [ 778.058351] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 778.058351] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 778.058351] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 778.058351] env[63202]: created_port_ids = self._update_ports_for_instance( [ 778.058351] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 778.058351] env[63202]: with excutils.save_and_reraise_exception(): [ 778.058351] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 778.058351] env[63202]: self.force_reraise() [ 778.058351] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 778.058351] env[63202]: raise self.value [ 778.058351] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 778.058351] env[63202]: updated_port = self._update_port( [ 778.058351] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 778.058351] env[63202]: _ensure_no_port_binding_failure(port) [ 778.058351] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 778.058351] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 778.059309] env[63202]: nova.exception.PortBindingFailed: Binding failed for port f6092551-11d9-42f1-bd5f-f010c6f57650, please check neutron logs for more information. [ 778.059309] env[63202]: Removing descriptor: 16 [ 778.073818] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.886s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.075030] env[63202]: ERROR nova.compute.manager [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0cdc9fcc-a493-46de-b498-4389665eaf90, please check neutron logs for more information. [ 778.075030] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] Traceback (most recent call last): [ 778.075030] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 778.075030] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] self.driver.spawn(context, instance, image_meta, [ 778.075030] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 778.075030] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 778.075030] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 778.075030] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] vm_ref = self.build_virtual_machine(instance, [ 778.075030] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 778.075030] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] vif_infos = vmwarevif.get_vif_info(self._session, [ 778.075030] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 778.075513] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] for vif in network_info: [ 778.075513] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 778.075513] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] return self._sync_wrapper(fn, *args, **kwargs) [ 778.075513] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 778.075513] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] self.wait() [ 778.075513] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 778.075513] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] self[:] = self._gt.wait() [ 778.075513] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 778.075513] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] return self._exit_event.wait() [ 778.075513] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 778.075513] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] result = hub.switch() [ 778.075513] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 778.075513] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] return self.greenlet.switch() [ 778.075921] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 778.075921] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] result = function(*args, **kwargs) [ 778.075921] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 778.075921] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] return func(*args, **kwargs) [ 778.075921] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 778.075921] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] raise e [ 778.075921] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 778.075921] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] nwinfo = self.network_api.allocate_for_instance( [ 778.075921] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 778.075921] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] created_port_ids = self._update_ports_for_instance( [ 778.075921] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 778.075921] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] with excutils.save_and_reraise_exception(): [ 778.075921] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 778.076327] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] self.force_reraise() [ 778.076327] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 778.076327] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] raise self.value [ 778.076327] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 778.076327] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] updated_port = self._update_port( [ 778.076327] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 778.076327] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] _ensure_no_port_binding_failure(port) [ 778.076327] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 778.076327] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] raise exception.PortBindingFailed(port_id=port['id']) [ 778.076327] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] nova.exception.PortBindingFailed: Binding failed for port 0cdc9fcc-a493-46de-b498-4389665eaf90, please check neutron logs for more information. [ 778.076327] env[63202]: ERROR nova.compute.manager [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] [ 778.076664] env[63202]: DEBUG nova.compute.utils [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] Binding failed for port 0cdc9fcc-a493-46de-b498-4389665eaf90, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 778.077244] env[63202]: DEBUG oslo_concurrency.lockutils [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.945s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 778.080438] env[63202]: DEBUG nova.compute.manager [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] Build of instance 611fead3-3d1b-41e4-9579-7ad1a1b754c3 was re-scheduled: Binding failed for port 0cdc9fcc-a493-46de-b498-4389665eaf90, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 778.080886] env[63202]: DEBUG nova.compute.manager [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 778.081221] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] Acquiring lock "refresh_cache-611fead3-3d1b-41e4-9579-7ad1a1b754c3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.081435] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] Acquired lock "refresh_cache-611fead3-3d1b-41e4-9579-7ad1a1b754c3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.081604] env[63202]: DEBUG nova.network.neutron [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 778.083925] env[63202]: INFO nova.compute.manager [-] [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] Took 1.03 seconds to deallocate network for instance. [ 778.085927] env[63202]: DEBUG nova.compute.claims [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 778.085927] env[63202]: DEBUG oslo_concurrency.lockutils [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 778.209696] env[63202]: DEBUG nova.compute.manager [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 778.236268] env[63202]: DEBUG nova.virt.hardware [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 778.236521] env[63202]: DEBUG nova.virt.hardware [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 778.237086] env[63202]: DEBUG nova.virt.hardware [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 778.237086] env[63202]: DEBUG nova.virt.hardware [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 778.237086] env[63202]: DEBUG nova.virt.hardware [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 778.237265] env[63202]: DEBUG nova.virt.hardware [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 778.237381] env[63202]: DEBUG nova.virt.hardware [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 778.237559] env[63202]: DEBUG nova.virt.hardware [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 778.237690] env[63202]: DEBUG nova.virt.hardware [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 778.237842] env[63202]: DEBUG nova.virt.hardware [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 778.238011] env[63202]: DEBUG nova.virt.hardware [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 778.238906] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5114b460-ea74-4314-a057-ec20ed9ed79b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.247233] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebf2483e-6982-4aff-94e0-6832d3628786 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.261426] env[63202]: ERROR nova.compute.manager [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f6092551-11d9-42f1-bd5f-f010c6f57650, please check neutron logs for more information. [ 778.261426] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] Traceback (most recent call last): [ 778.261426] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 778.261426] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] yield resources [ 778.261426] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 778.261426] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] self.driver.spawn(context, instance, image_meta, [ 778.261426] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 778.261426] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 778.261426] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 778.261426] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] vm_ref = self.build_virtual_machine(instance, [ 778.261426] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 778.261761] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] vif_infos = vmwarevif.get_vif_info(self._session, [ 778.261761] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 778.261761] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] for vif in network_info: [ 778.261761] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 778.261761] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] return self._sync_wrapper(fn, *args, **kwargs) [ 778.261761] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 778.261761] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] self.wait() [ 778.261761] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 778.261761] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] self[:] = self._gt.wait() [ 778.261761] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 778.261761] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] return self._exit_event.wait() [ 778.261761] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 778.261761] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] current.throw(*self._exc) [ 778.262228] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 778.262228] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] result = function(*args, **kwargs) [ 778.262228] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 778.262228] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] return func(*args, **kwargs) [ 778.262228] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 778.262228] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] raise e [ 778.262228] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 778.262228] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] nwinfo = self.network_api.allocate_for_instance( [ 778.262228] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 778.262228] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] created_port_ids = self._update_ports_for_instance( [ 778.262228] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 778.262228] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] with excutils.save_and_reraise_exception(): [ 778.262228] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 778.262638] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] self.force_reraise() [ 778.262638] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 778.262638] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] raise self.value [ 778.262638] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 778.262638] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] updated_port = self._update_port( [ 778.262638] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 778.262638] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] _ensure_no_port_binding_failure(port) [ 778.262638] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 778.262638] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] raise exception.PortBindingFailed(port_id=port['id']) [ 778.262638] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] nova.exception.PortBindingFailed: Binding failed for port f6092551-11d9-42f1-bd5f-f010c6f57650, please check neutron logs for more information. [ 778.262638] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] [ 778.262638] env[63202]: INFO nova.compute.manager [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] Terminating instance [ 778.263725] env[63202]: DEBUG oslo_concurrency.lockutils [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Acquiring lock "refresh_cache-c9e0ee31-de9e-420d-8dad-380391d6f4e3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.263852] env[63202]: DEBUG oslo_concurrency.lockutils [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Acquired lock "refresh_cache-c9e0ee31-de9e-420d-8dad-380391d6f4e3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.264023] env[63202]: DEBUG nova.network.neutron [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 778.349407] env[63202]: DEBUG oslo_concurrency.lockutils [req-914cf3a3-8195-4ba6-8ca7-097c9226a2eb req-168d95e7-4882-46b5-975b-0e556f6eb3ec service nova] Releasing lock "refresh_cache-e3a6ad78-4f46-42d4-935c-3cf310123530" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.349657] env[63202]: DEBUG nova.compute.manager [req-914cf3a3-8195-4ba6-8ca7-097c9226a2eb req-168d95e7-4882-46b5-975b-0e556f6eb3ec service nova] [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] Received event network-vif-deleted-ad906a1e-35dc-4bb7-85b6-2672972103d0 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 778.601185] env[63202]: DEBUG nova.network.neutron [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 778.719139] env[63202]: DEBUG nova.network.neutron [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.780641] env[63202]: DEBUG nova.network.neutron [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 778.877931] env[63202]: DEBUG nova.network.neutron [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.880355] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a75563b8-0635-4127-9fe1-86ca67ea1e99 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.890676] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1408620b-3a1a-4ee5-b714-963f84233b92 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.920833] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04d3524c-9e31-4413-837a-8cac320b3992 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.927975] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bdd8e4e-1250-40e4-ada9-141466f5f12d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.941920] env[63202]: DEBUG nova.compute.provider_tree [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 779.221717] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] Releasing lock "refresh_cache-611fead3-3d1b-41e4-9579-7ad1a1b754c3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.221916] env[63202]: DEBUG nova.compute.manager [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 779.222211] env[63202]: DEBUG nova.compute.manager [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 779.222421] env[63202]: DEBUG nova.network.neutron [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 779.236452] env[63202]: DEBUG nova.network.neutron [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 779.285418] env[63202]: DEBUG nova.compute.manager [req-e2ec1645-c403-4053-864a-8aa7a82b0aac req-4ea83fa5-a50b-4fa3-8f73-5a400a991252 service nova] [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] Received event network-changed-f6092551-11d9-42f1-bd5f-f010c6f57650 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 779.285663] env[63202]: DEBUG nova.compute.manager [req-e2ec1645-c403-4053-864a-8aa7a82b0aac req-4ea83fa5-a50b-4fa3-8f73-5a400a991252 service nova] [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] Refreshing instance network info cache due to event network-changed-f6092551-11d9-42f1-bd5f-f010c6f57650. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 779.285782] env[63202]: DEBUG oslo_concurrency.lockutils [req-e2ec1645-c403-4053-864a-8aa7a82b0aac req-4ea83fa5-a50b-4fa3-8f73-5a400a991252 service nova] Acquiring lock "refresh_cache-c9e0ee31-de9e-420d-8dad-380391d6f4e3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.384190] env[63202]: DEBUG oslo_concurrency.lockutils [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Releasing lock "refresh_cache-c9e0ee31-de9e-420d-8dad-380391d6f4e3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.384803] env[63202]: DEBUG nova.compute.manager [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 779.385071] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 779.385384] env[63202]: DEBUG oslo_concurrency.lockutils [req-e2ec1645-c403-4053-864a-8aa7a82b0aac req-4ea83fa5-a50b-4fa3-8f73-5a400a991252 service nova] Acquired lock "refresh_cache-c9e0ee31-de9e-420d-8dad-380391d6f4e3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.385574] env[63202]: DEBUG nova.network.neutron [req-e2ec1645-c403-4053-864a-8aa7a82b0aac req-4ea83fa5-a50b-4fa3-8f73-5a400a991252 service nova] [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] Refreshing network info cache for port f6092551-11d9-42f1-bd5f-f010c6f57650 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 779.386594] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-47cf7d0a-7ee0-46e0-8efd-7dd9b20e502b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.397370] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1db4c7e7-1a48-455c-9bab-9de7e6a33297 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.419808] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c9e0ee31-de9e-420d-8dad-380391d6f4e3 could not be found. [ 779.420069] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 779.420258] env[63202]: INFO nova.compute.manager [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] Took 0.04 seconds to destroy the instance on the hypervisor. [ 779.420494] env[63202]: DEBUG oslo.service.loopingcall [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 779.420732] env[63202]: DEBUG nova.compute.manager [-] [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 779.420821] env[63202]: DEBUG nova.network.neutron [-] [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 779.443106] env[63202]: DEBUG nova.network.neutron [-] [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 779.445023] env[63202]: DEBUG nova.scheduler.client.report [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 779.740244] env[63202]: DEBUG nova.network.neutron [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.905084] env[63202]: DEBUG nova.network.neutron [req-e2ec1645-c403-4053-864a-8aa7a82b0aac req-4ea83fa5-a50b-4fa3-8f73-5a400a991252 service nova] [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 779.950018] env[63202]: DEBUG nova.network.neutron [-] [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.951584] env[63202]: DEBUG oslo_concurrency.lockutils [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.874s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.952200] env[63202]: ERROR nova.compute.manager [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 35cca4a9-1e33-42f9-a7d4-50c5cea28d14, please check neutron logs for more information. [ 779.952200] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] Traceback (most recent call last): [ 779.952200] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 779.952200] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] self.driver.spawn(context, instance, image_meta, [ 779.952200] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 779.952200] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 779.952200] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 779.952200] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] vm_ref = self.build_virtual_machine(instance, [ 779.952200] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 779.952200] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] vif_infos = vmwarevif.get_vif_info(self._session, [ 779.952200] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 779.952637] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] for vif in network_info: [ 779.952637] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 779.952637] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] return self._sync_wrapper(fn, *args, **kwargs) [ 779.952637] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 779.952637] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] self.wait() [ 779.952637] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 779.952637] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] self[:] = self._gt.wait() [ 779.952637] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 779.952637] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] return self._exit_event.wait() [ 779.952637] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 779.952637] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] current.throw(*self._exc) [ 779.952637] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 779.952637] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] result = function(*args, **kwargs) [ 779.953044] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 779.953044] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] return func(*args, **kwargs) [ 779.953044] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 779.953044] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] raise e [ 779.953044] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 779.953044] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] nwinfo = self.network_api.allocate_for_instance( [ 779.953044] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 779.953044] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] created_port_ids = self._update_ports_for_instance( [ 779.953044] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 779.953044] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] with excutils.save_and_reraise_exception(): [ 779.953044] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 779.953044] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] self.force_reraise() [ 779.953044] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 779.953481] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] raise self.value [ 779.953481] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 779.953481] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] updated_port = self._update_port( [ 779.953481] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 779.953481] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] _ensure_no_port_binding_failure(port) [ 779.953481] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 779.953481] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] raise exception.PortBindingFailed(port_id=port['id']) [ 779.953481] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] nova.exception.PortBindingFailed: Binding failed for port 35cca4a9-1e33-42f9-a7d4-50c5cea28d14, please check neutron logs for more information. [ 779.953481] env[63202]: ERROR nova.compute.manager [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] [ 779.953481] env[63202]: DEBUG nova.compute.utils [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] Binding failed for port 35cca4a9-1e33-42f9-a7d4-50c5cea28d14, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 779.954177] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 16.879s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.954249] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.954392] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63202) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 779.954676] env[63202]: DEBUG oslo_concurrency.lockutils [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.366s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.957799] env[63202]: DEBUG nova.compute.manager [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] Build of instance 4cd7f7b3-d947-4745-8fd7-940076865e3b was re-scheduled: Binding failed for port 35cca4a9-1e33-42f9-a7d4-50c5cea28d14, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 779.958239] env[63202]: DEBUG nova.compute.manager [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 779.958504] env[63202]: DEBUG oslo_concurrency.lockutils [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Acquiring lock "refresh_cache-4cd7f7b3-d947-4745-8fd7-940076865e3b" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.958657] env[63202]: DEBUG oslo_concurrency.lockutils [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Acquired lock "refresh_cache-4cd7f7b3-d947-4745-8fd7-940076865e3b" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.958813] env[63202]: DEBUG nova.network.neutron [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 779.960381] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a65ee785-6db1-411b-80fe-f1dcdc088a38 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.969378] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33a56d4f-3e00-43ba-8936-801478fad0f3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.984770] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1031d129-ab7e-4eef-89dc-1f6529488586 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.991406] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-724d78f6-d3ed-4fcb-8fbc-1577cd4415a3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.022349] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181171MB free_disk=176GB free_vcpus=48 pci_devices=None {{(pid=63202) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 780.022503] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 780.024845] env[63202]: DEBUG nova.network.neutron [req-e2ec1645-c403-4053-864a-8aa7a82b0aac req-4ea83fa5-a50b-4fa3-8f73-5a400a991252 service nova] [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.242682] env[63202]: INFO nova.compute.manager [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] [instance: 611fead3-3d1b-41e4-9579-7ad1a1b754c3] Took 1.02 seconds to deallocate network for instance. [ 780.455473] env[63202]: INFO nova.compute.manager [-] [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] Took 1.03 seconds to deallocate network for instance. [ 780.457792] env[63202]: DEBUG nova.compute.claims [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 780.457976] env[63202]: DEBUG oslo_concurrency.lockutils [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 780.479699] env[63202]: DEBUG nova.network.neutron [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 780.526380] env[63202]: DEBUG oslo_concurrency.lockutils [req-e2ec1645-c403-4053-864a-8aa7a82b0aac req-4ea83fa5-a50b-4fa3-8f73-5a400a991252 service nova] Releasing lock "refresh_cache-c9e0ee31-de9e-420d-8dad-380391d6f4e3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.526566] env[63202]: DEBUG nova.compute.manager [req-e2ec1645-c403-4053-864a-8aa7a82b0aac req-4ea83fa5-a50b-4fa3-8f73-5a400a991252 service nova] [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] Received event network-vif-deleted-f6092551-11d9-42f1-bd5f-f010c6f57650 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 780.552159] env[63202]: DEBUG nova.network.neutron [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.729183] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91f1e48d-29e0-4127-8623-09f06073131c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.737216] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78306a09-dec9-4fb1-a0f3-3a48038ce452 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.770551] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08e3d711-a9d7-4c2b-9249-b7bd792bc366 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.778800] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d34e0a8-dbf3-474a-a5af-aa27ff533605 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.793395] env[63202]: DEBUG nova.compute.provider_tree [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 781.055562] env[63202]: DEBUG oslo_concurrency.lockutils [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Releasing lock "refresh_cache-4cd7f7b3-d947-4745-8fd7-940076865e3b" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.055790] env[63202]: DEBUG nova.compute.manager [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 781.055965] env[63202]: DEBUG nova.compute.manager [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 781.056144] env[63202]: DEBUG nova.network.neutron [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 781.086465] env[63202]: DEBUG nova.network.neutron [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 781.295728] env[63202]: INFO nova.scheduler.client.report [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] Deleted allocations for instance 611fead3-3d1b-41e4-9579-7ad1a1b754c3 [ 781.302360] env[63202]: DEBUG nova.scheduler.client.report [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 781.589328] env[63202]: DEBUG nova.network.neutron [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.807961] env[63202]: DEBUG oslo_concurrency.lockutils [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.853s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.809119] env[63202]: ERROR nova.compute.manager [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a2c4847c-da1f-4c90-87f0-4678eeaf1504, please check neutron logs for more information. [ 781.809119] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] Traceback (most recent call last): [ 781.809119] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 781.809119] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] self.driver.spawn(context, instance, image_meta, [ 781.809119] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 781.809119] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] self._vmops.spawn(context, instance, image_meta, injected_files, [ 781.809119] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 781.809119] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] vm_ref = self.build_virtual_machine(instance, [ 781.809119] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 781.809119] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] vif_infos = vmwarevif.get_vif_info(self._session, [ 781.809119] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 781.809507] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] for vif in network_info: [ 781.809507] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 781.809507] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] return self._sync_wrapper(fn, *args, **kwargs) [ 781.809507] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 781.809507] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] self.wait() [ 781.809507] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 781.809507] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] self[:] = self._gt.wait() [ 781.809507] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 781.809507] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] return self._exit_event.wait() [ 781.809507] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 781.809507] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] result = hub.switch() [ 781.809507] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 781.809507] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] return self.greenlet.switch() [ 781.809908] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 781.809908] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] result = function(*args, **kwargs) [ 781.809908] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 781.809908] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] return func(*args, **kwargs) [ 781.809908] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 781.809908] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] raise e [ 781.809908] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 781.809908] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] nwinfo = self.network_api.allocate_for_instance( [ 781.809908] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 781.809908] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] created_port_ids = self._update_ports_for_instance( [ 781.809908] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 781.809908] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] with excutils.save_and_reraise_exception(): [ 781.809908] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 781.810335] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] self.force_reraise() [ 781.810335] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 781.810335] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] raise self.value [ 781.810335] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 781.810335] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] updated_port = self._update_port( [ 781.810335] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 781.810335] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] _ensure_no_port_binding_failure(port) [ 781.810335] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 781.810335] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] raise exception.PortBindingFailed(port_id=port['id']) [ 781.810335] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] nova.exception.PortBindingFailed: Binding failed for port a2c4847c-da1f-4c90-87f0-4678eeaf1504, please check neutron logs for more information. [ 781.810335] env[63202]: ERROR nova.compute.manager [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] [ 781.810668] env[63202]: DEBUG nova.compute.utils [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] Binding failed for port a2c4847c-da1f-4c90-87f0-4678eeaf1504, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 781.811115] env[63202]: DEBUG oslo_concurrency.lockutils [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.718s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.813881] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6ad886d4-29c3-4941-8a6d-834df77a466f tempest-InstanceActionsNegativeTestJSON-468774283 tempest-InstanceActionsNegativeTestJSON-468774283-project-member] Lock "611fead3-3d1b-41e4-9579-7ad1a1b754c3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 143.028s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.814303] env[63202]: DEBUG nova.compute.manager [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] Build of instance ba2d047d-d30f-4ba0-bcfb-787c5a3ae516 was re-scheduled: Binding failed for port a2c4847c-da1f-4c90-87f0-4678eeaf1504, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 781.814719] env[63202]: DEBUG nova.compute.manager [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 781.814953] env[63202]: DEBUG oslo_concurrency.lockutils [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Acquiring lock "refresh_cache-ba2d047d-d30f-4ba0-bcfb-787c5a3ae516" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.815124] env[63202]: DEBUG oslo_concurrency.lockutils [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Acquired lock "refresh_cache-ba2d047d-d30f-4ba0-bcfb-787c5a3ae516" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.815294] env[63202]: DEBUG nova.network.neutron [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 782.092539] env[63202]: INFO nova.compute.manager [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: 4cd7f7b3-d947-4745-8fd7-940076865e3b] Took 1.04 seconds to deallocate network for instance. [ 782.322190] env[63202]: DEBUG nova.compute.manager [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 782.338536] env[63202]: DEBUG nova.network.neutron [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 782.419095] env[63202]: DEBUG nova.network.neutron [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.670588] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1c59742-66cb-4a21-b31d-e84270bc6d83 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.681182] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-202c66ab-8762-4eaa-be44-f5325ff2266e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.713344] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-816e9446-8758-41e6-b70b-f46f1998e44c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.721076] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23157f6b-318e-4635-9cdc-ae5947cda935 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.736351] env[63202]: DEBUG nova.compute.provider_tree [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 782.841196] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.927630] env[63202]: DEBUG oslo_concurrency.lockutils [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Releasing lock "refresh_cache-ba2d047d-d30f-4ba0-bcfb-787c5a3ae516" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.927874] env[63202]: DEBUG nova.compute.manager [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 782.928064] env[63202]: DEBUG nova.compute.manager [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 782.928264] env[63202]: DEBUG nova.network.neutron [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 782.947773] env[63202]: DEBUG nova.network.neutron [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 783.121906] env[63202]: INFO nova.scheduler.client.report [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Deleted allocations for instance 4cd7f7b3-d947-4745-8fd7-940076865e3b [ 783.239752] env[63202]: DEBUG nova.scheduler.client.report [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 783.450660] env[63202]: DEBUG nova.network.neutron [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.633030] env[63202]: DEBUG oslo_concurrency.lockutils [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Lock "4cd7f7b3-d947-4745-8fd7-940076865e3b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 141.646s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.744731] env[63202]: DEBUG oslo_concurrency.lockutils [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.934s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.745400] env[63202]: ERROR nova.compute.manager [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port cc4ef3ae-fbb2-4f70-b9e7-103b29e5b9f6, please check neutron logs for more information. [ 783.745400] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] Traceback (most recent call last): [ 783.745400] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 783.745400] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] self.driver.spawn(context, instance, image_meta, [ 783.745400] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 783.745400] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 783.745400] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 783.745400] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] vm_ref = self.build_virtual_machine(instance, [ 783.745400] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 783.745400] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] vif_infos = vmwarevif.get_vif_info(self._session, [ 783.745400] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 783.745838] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] for vif in network_info: [ 783.745838] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 783.745838] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] return self._sync_wrapper(fn, *args, **kwargs) [ 783.745838] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 783.745838] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] self.wait() [ 783.745838] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 783.745838] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] self[:] = self._gt.wait() [ 783.745838] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 783.745838] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] return self._exit_event.wait() [ 783.745838] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 783.745838] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] result = hub.switch() [ 783.745838] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 783.745838] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] return self.greenlet.switch() [ 783.746271] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 783.746271] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] result = function(*args, **kwargs) [ 783.746271] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 783.746271] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] return func(*args, **kwargs) [ 783.746271] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 783.746271] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] raise e [ 783.746271] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 783.746271] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] nwinfo = self.network_api.allocate_for_instance( [ 783.746271] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 783.746271] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] created_port_ids = self._update_ports_for_instance( [ 783.746271] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 783.746271] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] with excutils.save_and_reraise_exception(): [ 783.746271] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 783.746776] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] self.force_reraise() [ 783.746776] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 783.746776] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] raise self.value [ 783.746776] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 783.746776] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] updated_port = self._update_port( [ 783.746776] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 783.746776] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] _ensure_no_port_binding_failure(port) [ 783.746776] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 783.746776] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] raise exception.PortBindingFailed(port_id=port['id']) [ 783.746776] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] nova.exception.PortBindingFailed: Binding failed for port cc4ef3ae-fbb2-4f70-b9e7-103b29e5b9f6, please check neutron logs for more information. [ 783.746776] env[63202]: ERROR nova.compute.manager [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] [ 783.747215] env[63202]: DEBUG nova.compute.utils [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] Binding failed for port cc4ef3ae-fbb2-4f70-b9e7-103b29e5b9f6, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 783.747322] env[63202]: DEBUG oslo_concurrency.lockutils [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.384s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.750122] env[63202]: DEBUG nova.compute.manager [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] Build of instance 1ac8e024-65d1-4250-a946-858aaff852c8 was re-scheduled: Binding failed for port cc4ef3ae-fbb2-4f70-b9e7-103b29e5b9f6, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 783.750547] env[63202]: DEBUG nova.compute.manager [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 783.750767] env[63202]: DEBUG oslo_concurrency.lockutils [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Acquiring lock "refresh_cache-1ac8e024-65d1-4250-a946-858aaff852c8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.750910] env[63202]: DEBUG oslo_concurrency.lockutils [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Acquired lock "refresh_cache-1ac8e024-65d1-4250-a946-858aaff852c8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.751108] env[63202]: DEBUG nova.network.neutron [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 783.957716] env[63202]: INFO nova.compute.manager [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] [instance: ba2d047d-d30f-4ba0-bcfb-787c5a3ae516] Took 1.03 seconds to deallocate network for instance. [ 784.135561] env[63202]: DEBUG nova.compute.manager [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 784.287483] env[63202]: DEBUG nova.network.neutron [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 784.338045] env[63202]: DEBUG nova.network.neutron [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.522361] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0af76116-ed10-4b26-ac92-b7c0c3175572 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.530792] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12d7501c-0844-4a52-899e-780df59a2c16 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.560854] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be6faf64-b093-43cd-ba26-f15d8497f9e1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.569309] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3d665d3-0d01-4b32-8366-48b895460fcb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.583917] env[63202]: DEBUG nova.compute.provider_tree [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 784.655883] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.842992] env[63202]: DEBUG oslo_concurrency.lockutils [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Releasing lock "refresh_cache-1ac8e024-65d1-4250-a946-858aaff852c8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.842992] env[63202]: DEBUG nova.compute.manager [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 784.842992] env[63202]: DEBUG nova.compute.manager [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 784.842992] env[63202]: DEBUG nova.network.neutron [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 784.876690] env[63202]: DEBUG nova.network.neutron [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 784.987024] env[63202]: INFO nova.scheduler.client.report [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Deleted allocations for instance ba2d047d-d30f-4ba0-bcfb-787c5a3ae516 [ 785.088031] env[63202]: DEBUG nova.scheduler.client.report [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 785.380300] env[63202]: DEBUG nova.network.neutron [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.496511] env[63202]: DEBUG oslo_concurrency.lockutils [None req-25d69db4-7427-4d39-ab38-eae852fbae1a tempest-MultipleCreateTestJSON-2008172540 tempest-MultipleCreateTestJSON-2008172540-project-member] Lock "ba2d047d-d30f-4ba0-bcfb-787c5a3ae516" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 143.472s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.592312] env[63202]: DEBUG oslo_concurrency.lockutils [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.845s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.592936] env[63202]: ERROR nova.compute.manager [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3c2c28dc-66c3-479c-a3a2-2d1ed7c3ae4d, please check neutron logs for more information. [ 785.592936] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] Traceback (most recent call last): [ 785.592936] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 785.592936] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] self.driver.spawn(context, instance, image_meta, [ 785.592936] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 785.592936] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] self._vmops.spawn(context, instance, image_meta, injected_files, [ 785.592936] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 785.592936] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] vm_ref = self.build_virtual_machine(instance, [ 785.592936] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 785.592936] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] vif_infos = vmwarevif.get_vif_info(self._session, [ 785.592936] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 785.594529] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] for vif in network_info: [ 785.594529] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 785.594529] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] return self._sync_wrapper(fn, *args, **kwargs) [ 785.594529] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 785.594529] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] self.wait() [ 785.594529] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 785.594529] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] self[:] = self._gt.wait() [ 785.594529] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 785.594529] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] return self._exit_event.wait() [ 785.594529] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 785.594529] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] current.throw(*self._exc) [ 785.594529] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 785.594529] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] result = function(*args, **kwargs) [ 785.595211] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 785.595211] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] return func(*args, **kwargs) [ 785.595211] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 785.595211] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] raise e [ 785.595211] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 785.595211] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] nwinfo = self.network_api.allocate_for_instance( [ 785.595211] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 785.595211] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] created_port_ids = self._update_ports_for_instance( [ 785.595211] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 785.595211] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] with excutils.save_and_reraise_exception(): [ 785.595211] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 785.595211] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] self.force_reraise() [ 785.595211] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 785.595602] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] raise self.value [ 785.595602] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 785.595602] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] updated_port = self._update_port( [ 785.595602] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 785.595602] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] _ensure_no_port_binding_failure(port) [ 785.595602] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 785.595602] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] raise exception.PortBindingFailed(port_id=port['id']) [ 785.595602] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] nova.exception.PortBindingFailed: Binding failed for port 3c2c28dc-66c3-479c-a3a2-2d1ed7c3ae4d, please check neutron logs for more information. [ 785.595602] env[63202]: ERROR nova.compute.manager [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] [ 785.595602] env[63202]: DEBUG nova.compute.utils [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] Binding failed for port 3c2c28dc-66c3-479c-a3a2-2d1ed7c3ae4d, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 785.595868] env[63202]: DEBUG nova.compute.manager [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] Build of instance 3080314c-938e-4c27-bffd-547bdc7e6e38 was re-scheduled: Binding failed for port 3c2c28dc-66c3-479c-a3a2-2d1ed7c3ae4d, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 785.598643] env[63202]: DEBUG nova.compute.manager [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 785.598643] env[63202]: DEBUG oslo_concurrency.lockutils [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Acquiring lock "refresh_cache-3080314c-938e-4c27-bffd-547bdc7e6e38" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 785.598643] env[63202]: DEBUG oslo_concurrency.lockutils [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Acquired lock "refresh_cache-3080314c-938e-4c27-bffd-547bdc7e6e38" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 785.598643] env[63202]: DEBUG nova.network.neutron [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 785.599832] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.507s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.885725] env[63202]: INFO nova.compute.manager [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 1ac8e024-65d1-4250-a946-858aaff852c8] Took 1.04 seconds to deallocate network for instance. [ 786.001056] env[63202]: DEBUG nova.compute.manager [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 786.134020] env[63202]: DEBUG nova.network.neutron [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 786.263636] env[63202]: DEBUG nova.network.neutron [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.417201] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dc60bcf-2c38-43b7-a663-1d5bb9953c34 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.425108] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cf00e0e-d7b7-4b1b-bc87-9f1b389d30ca {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.465140] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8a3215a-d4a5-4f71-926d-c1f92a613862 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.473099] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39284c71-d898-47b7-9812-3f6a078f7989 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.491664] env[63202]: DEBUG nova.compute.provider_tree [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 786.525341] env[63202]: DEBUG oslo_concurrency.lockutils [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.765723] env[63202]: DEBUG oslo_concurrency.lockutils [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Releasing lock "refresh_cache-3080314c-938e-4c27-bffd-547bdc7e6e38" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 786.765950] env[63202]: DEBUG nova.compute.manager [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 786.766123] env[63202]: DEBUG nova.compute.manager [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 786.766283] env[63202]: DEBUG nova.network.neutron [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 786.782718] env[63202]: DEBUG nova.network.neutron [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 786.916965] env[63202]: INFO nova.scheduler.client.report [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Deleted allocations for instance 1ac8e024-65d1-4250-a946-858aaff852c8 [ 786.996375] env[63202]: DEBUG nova.scheduler.client.report [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 787.168451] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Acquiring lock "cedc3a06-2123-4c5e-a6c3-599a3efc3c65" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.168768] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Lock "cedc3a06-2123-4c5e-a6c3-599a3efc3c65" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.285708] env[63202]: DEBUG nova.network.neutron [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 787.427208] env[63202]: DEBUG oslo_concurrency.lockutils [None req-fa248bb5-9c3d-4f48-ba04-b18af933bbc8 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Lock "1ac8e024-65d1-4250-a946-858aaff852c8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 144.482s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.501897] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.903s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.502505] env[63202]: ERROR nova.compute.manager [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 52847cc7-89c1-4996-92bd-ec6cefb47d70, please check neutron logs for more information. [ 787.502505] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Traceback (most recent call last): [ 787.502505] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 787.502505] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] self.driver.spawn(context, instance, image_meta, [ 787.502505] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 787.502505] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 787.502505] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 787.502505] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] vm_ref = self.build_virtual_machine(instance, [ 787.502505] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 787.502505] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] vif_infos = vmwarevif.get_vif_info(self._session, [ 787.502505] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 787.502937] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] for vif in network_info: [ 787.502937] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 787.502937] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] return self._sync_wrapper(fn, *args, **kwargs) [ 787.502937] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 787.502937] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] self.wait() [ 787.502937] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 787.502937] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] self[:] = self._gt.wait() [ 787.502937] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 787.502937] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] return self._exit_event.wait() [ 787.502937] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 787.502937] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] current.throw(*self._exc) [ 787.502937] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 787.502937] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] result = function(*args, **kwargs) [ 787.503408] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 787.503408] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] return func(*args, **kwargs) [ 787.503408] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 787.503408] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] raise e [ 787.503408] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 787.503408] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] nwinfo = self.network_api.allocate_for_instance( [ 787.503408] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 787.503408] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] created_port_ids = self._update_ports_for_instance( [ 787.503408] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 787.503408] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] with excutils.save_and_reraise_exception(): [ 787.503408] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 787.503408] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] self.force_reraise() [ 787.503408] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 787.503836] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] raise self.value [ 787.503836] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 787.503836] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] updated_port = self._update_port( [ 787.503836] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 787.503836] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] _ensure_no_port_binding_failure(port) [ 787.503836] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 787.503836] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] raise exception.PortBindingFailed(port_id=port['id']) [ 787.503836] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] nova.exception.PortBindingFailed: Binding failed for port 52847cc7-89c1-4996-92bd-ec6cefb47d70, please check neutron logs for more information. [ 787.503836] env[63202]: ERROR nova.compute.manager [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] [ 787.503836] env[63202]: DEBUG nova.compute.utils [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Binding failed for port 52847cc7-89c1-4996-92bd-ec6cefb47d70, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 787.505394] env[63202]: DEBUG oslo_concurrency.lockutils [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.205s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.507224] env[63202]: DEBUG nova.compute.manager [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Build of instance 7efc454e-337d-43db-9076-bfc5b89eeea4 was re-scheduled: Binding failed for port 52847cc7-89c1-4996-92bd-ec6cefb47d70, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 787.507652] env[63202]: DEBUG nova.compute.manager [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 787.507868] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] Acquiring lock "refresh_cache-7efc454e-337d-43db-9076-bfc5b89eeea4" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.508015] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] Acquired lock "refresh_cache-7efc454e-337d-43db-9076-bfc5b89eeea4" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.511796] env[63202]: DEBUG nova.network.neutron [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 787.788831] env[63202]: INFO nova.compute.manager [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] [instance: 3080314c-938e-4c27-bffd-547bdc7e6e38] Took 1.02 seconds to deallocate network for instance. [ 787.929843] env[63202]: DEBUG nova.compute.manager [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 788.036166] env[63202]: DEBUG nova.network.neutron [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 788.134484] env[63202]: DEBUG nova.network.neutron [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.278124] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48fa21bd-ab81-4424-bc0e-e2224eb025df {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.286557] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4ddb231-d217-4bd4-b099-628ac17d5d62 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.319029] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a758ebe-6536-40c0-8227-8c1af2702480 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.326889] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09500845-0be0-497b-bf86-da5a2e408901 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.341033] env[63202]: DEBUG nova.compute.provider_tree [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 788.449748] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.637371] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] Releasing lock "refresh_cache-7efc454e-337d-43db-9076-bfc5b89eeea4" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.637604] env[63202]: DEBUG nova.compute.manager [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 788.637780] env[63202]: DEBUG nova.compute.manager [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 788.637945] env[63202]: DEBUG nova.network.neutron [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 788.662961] env[63202]: DEBUG nova.network.neutron [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 788.846980] env[63202]: DEBUG nova.scheduler.client.report [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 788.850637] env[63202]: INFO nova.scheduler.client.report [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Deleted allocations for instance 3080314c-938e-4c27-bffd-547bdc7e6e38 [ 789.166013] env[63202]: DEBUG nova.network.neutron [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.349764] env[63202]: DEBUG oslo_concurrency.lockutils [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.845s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.350414] env[63202]: ERROR nova.compute.manager [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: fea91da2-186a-44b6-8e20-535e3b123890] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 58102740-cab3-4fc7-8dbf-725c4f0131f6, please check neutron logs for more information. [ 789.350414] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] Traceback (most recent call last): [ 789.350414] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 789.350414] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] self.driver.spawn(context, instance, image_meta, [ 789.350414] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 789.350414] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] self._vmops.spawn(context, instance, image_meta, injected_files, [ 789.350414] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 789.350414] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] vm_ref = self.build_virtual_machine(instance, [ 789.350414] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 789.350414] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] vif_infos = vmwarevif.get_vif_info(self._session, [ 789.350414] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 789.350785] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] for vif in network_info: [ 789.350785] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 789.350785] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] return self._sync_wrapper(fn, *args, **kwargs) [ 789.350785] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 789.350785] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] self.wait() [ 789.350785] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 789.350785] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] self[:] = self._gt.wait() [ 789.350785] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 789.350785] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] return self._exit_event.wait() [ 789.350785] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 789.350785] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] result = hub.switch() [ 789.350785] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 789.350785] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] return self.greenlet.switch() [ 789.351207] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 789.351207] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] result = function(*args, **kwargs) [ 789.351207] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 789.351207] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] return func(*args, **kwargs) [ 789.351207] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 789.351207] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] raise e [ 789.351207] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 789.351207] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] nwinfo = self.network_api.allocate_for_instance( [ 789.351207] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 789.351207] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] created_port_ids = self._update_ports_for_instance( [ 789.351207] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 789.351207] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] with excutils.save_and_reraise_exception(): [ 789.351207] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 789.351603] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] self.force_reraise() [ 789.351603] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 789.351603] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] raise self.value [ 789.351603] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 789.351603] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] updated_port = self._update_port( [ 789.351603] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 789.351603] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] _ensure_no_port_binding_failure(port) [ 789.351603] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 789.351603] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] raise exception.PortBindingFailed(port_id=port['id']) [ 789.351603] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] nova.exception.PortBindingFailed: Binding failed for port 58102740-cab3-4fc7-8dbf-725c4f0131f6, please check neutron logs for more information. [ 789.351603] env[63202]: ERROR nova.compute.manager [instance: fea91da2-186a-44b6-8e20-535e3b123890] [ 789.351960] env[63202]: DEBUG nova.compute.utils [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: fea91da2-186a-44b6-8e20-535e3b123890] Binding failed for port 58102740-cab3-4fc7-8dbf-725c4f0131f6, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 789.352677] env[63202]: DEBUG oslo_concurrency.lockutils [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.884s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.355439] env[63202]: DEBUG nova.compute.manager [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: fea91da2-186a-44b6-8e20-535e3b123890] Build of instance fea91da2-186a-44b6-8e20-535e3b123890 was re-scheduled: Binding failed for port 58102740-cab3-4fc7-8dbf-725c4f0131f6, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 789.355852] env[63202]: DEBUG nova.compute.manager [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: fea91da2-186a-44b6-8e20-535e3b123890] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 789.356088] env[63202]: DEBUG oslo_concurrency.lockutils [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Acquiring lock "refresh_cache-fea91da2-186a-44b6-8e20-535e3b123890" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 789.356235] env[63202]: DEBUG oslo_concurrency.lockutils [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Acquired lock "refresh_cache-fea91da2-186a-44b6-8e20-535e3b123890" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.356390] env[63202]: DEBUG nova.network.neutron [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: fea91da2-186a-44b6-8e20-535e3b123890] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 789.358512] env[63202]: DEBUG oslo_concurrency.lockutils [None req-38baac18-a0dd-44cc-84f0-5e1c6b401767 tempest-ListImageFiltersTestJSON-27330218 tempest-ListImageFiltersTestJSON-27330218-project-member] Lock "3080314c-938e-4c27-bffd-547bdc7e6e38" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 145.783s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.668808] env[63202]: INFO nova.compute.manager [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] [instance: 7efc454e-337d-43db-9076-bfc5b89eeea4] Took 1.03 seconds to deallocate network for instance. [ 789.863213] env[63202]: DEBUG nova.compute.manager [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 789.916156] env[63202]: DEBUG nova.network.neutron [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: fea91da2-186a-44b6-8e20-535e3b123890] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 790.026365] env[63202]: DEBUG nova.network.neutron [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: fea91da2-186a-44b6-8e20-535e3b123890] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.167929] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9035f217-f9f4-4118-a951-0c2996495c34 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.180605] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31e5a216-7644-4223-a107-8e91aea21416 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.216613] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08053ffe-75ad-4931-98b9-ab07afd17c94 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.226587] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12a8f68f-88b4-4995-a8fe-fe2bd7eaeab4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.242021] env[63202]: DEBUG nova.compute.provider_tree [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 790.389373] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.528957] env[63202]: DEBUG oslo_concurrency.lockutils [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Releasing lock "refresh_cache-fea91da2-186a-44b6-8e20-535e3b123890" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.530357] env[63202]: DEBUG nova.compute.manager [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 790.530482] env[63202]: DEBUG nova.compute.manager [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: fea91da2-186a-44b6-8e20-535e3b123890] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 790.530632] env[63202]: DEBUG nova.network.neutron [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: fea91da2-186a-44b6-8e20-535e3b123890] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 790.546224] env[63202]: DEBUG nova.network.neutron [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: fea91da2-186a-44b6-8e20-535e3b123890] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 790.713053] env[63202]: INFO nova.scheduler.client.report [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] Deleted allocations for instance 7efc454e-337d-43db-9076-bfc5b89eeea4 [ 790.746031] env[63202]: DEBUG nova.scheduler.client.report [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 791.050120] env[63202]: DEBUG nova.network.neutron [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: fea91da2-186a-44b6-8e20-535e3b123890] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.225451] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d698b416-ac21-43af-ab61-11bf421bd454 tempest-ServersTestBootFromVolume-1982153917 tempest-ServersTestBootFromVolume-1982153917-project-member] Lock "7efc454e-337d-43db-9076-bfc5b89eeea4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 146.994s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.254117] env[63202]: DEBUG oslo_concurrency.lockutils [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.900s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.254117] env[63202]: ERROR nova.compute.manager [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0ab92838-e88e-4996-94b8-21a3a09753d8, please check neutron logs for more information. [ 791.254117] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] Traceback (most recent call last): [ 791.254117] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 791.254117] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] self.driver.spawn(context, instance, image_meta, [ 791.254117] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 791.254117] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 791.254117] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 791.254117] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] vm_ref = self.build_virtual_machine(instance, [ 791.254491] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 791.254491] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] vif_infos = vmwarevif.get_vif_info(self._session, [ 791.254491] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 791.254491] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] for vif in network_info: [ 791.254491] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 791.254491] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] return self._sync_wrapper(fn, *args, **kwargs) [ 791.254491] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 791.254491] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] self.wait() [ 791.254491] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 791.254491] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] self[:] = self._gt.wait() [ 791.254491] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 791.254491] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] return self._exit_event.wait() [ 791.254491] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 791.254897] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] current.throw(*self._exc) [ 791.254897] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 791.254897] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] result = function(*args, **kwargs) [ 791.254897] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 791.254897] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] return func(*args, **kwargs) [ 791.254897] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 791.254897] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] raise e [ 791.254897] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 791.254897] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] nwinfo = self.network_api.allocate_for_instance( [ 791.254897] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 791.254897] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] created_port_ids = self._update_ports_for_instance( [ 791.254897] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 791.254897] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] with excutils.save_and_reraise_exception(): [ 791.256407] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 791.256407] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] self.force_reraise() [ 791.256407] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 791.256407] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] raise self.value [ 791.256407] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 791.256407] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] updated_port = self._update_port( [ 791.256407] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 791.256407] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] _ensure_no_port_binding_failure(port) [ 791.256407] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 791.256407] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] raise exception.PortBindingFailed(port_id=port['id']) [ 791.256407] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] nova.exception.PortBindingFailed: Binding failed for port 0ab92838-e88e-4996-94b8-21a3a09753d8, please check neutron logs for more information. [ 791.256407] env[63202]: ERROR nova.compute.manager [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] [ 791.256831] env[63202]: DEBUG nova.compute.utils [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] Binding failed for port 0ab92838-e88e-4996-94b8-21a3a09753d8, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 791.256831] env[63202]: DEBUG oslo_concurrency.lockutils [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.170s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.259065] env[63202]: DEBUG nova.compute.manager [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] Build of instance 8045881e-9bce-46e7-98c6-a7989f61a31e was re-scheduled: Binding failed for port 0ab92838-e88e-4996-94b8-21a3a09753d8, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 791.259492] env[63202]: DEBUG nova.compute.manager [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 791.259829] env[63202]: DEBUG oslo_concurrency.lockutils [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "refresh_cache-8045881e-9bce-46e7-98c6-a7989f61a31e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.259829] env[63202]: DEBUG oslo_concurrency.lockutils [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquired lock "refresh_cache-8045881e-9bce-46e7-98c6-a7989f61a31e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.260018] env[63202]: DEBUG nova.network.neutron [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 791.557461] env[63202]: INFO nova.compute.manager [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: fea91da2-186a-44b6-8e20-535e3b123890] Took 1.02 seconds to deallocate network for instance. [ 791.729547] env[63202]: DEBUG nova.compute.manager [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 791.794805] env[63202]: DEBUG nova.network.neutron [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 791.930142] env[63202]: DEBUG nova.network.neutron [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.069654] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4109e01-e0fc-4504-b581-6820876d0a2e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.082579] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0186cd9-ee08-4fe0-bb62-ba5534c2467c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.122963] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c89b563f-3805-4629-9621-322b0c44243f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.133197] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60ac1ea6-11dc-4fd7-bf26-76a8d2294e7e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.148208] env[63202]: DEBUG nova.compute.provider_tree [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 792.258863] env[63202]: DEBUG oslo_concurrency.lockutils [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.433828] env[63202]: DEBUG oslo_concurrency.lockutils [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Releasing lock "refresh_cache-8045881e-9bce-46e7-98c6-a7989f61a31e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.434085] env[63202]: DEBUG nova.compute.manager [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 792.434319] env[63202]: DEBUG nova.compute.manager [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 792.434518] env[63202]: DEBUG nova.network.neutron [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 792.465288] env[63202]: DEBUG nova.network.neutron [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 792.596232] env[63202]: INFO nova.scheduler.client.report [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Deleted allocations for instance fea91da2-186a-44b6-8e20-535e3b123890 [ 792.653565] env[63202]: DEBUG nova.scheduler.client.report [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 792.722604] env[63202]: DEBUG oslo_concurrency.lockutils [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Acquiring lock "0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.722841] env[63202]: DEBUG oslo_concurrency.lockutils [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lock "0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.759273] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Acquiring lock "a800fa5d-46a1-4a96-8ea2-26e34b869aa6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.759506] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Lock "a800fa5d-46a1-4a96-8ea2-26e34b869aa6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.968641] env[63202]: DEBUG nova.network.neutron [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.104514] env[63202]: DEBUG oslo_concurrency.lockutils [None req-64215076-0610-463c-a54e-010c51d42f29 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Lock "fea91da2-186a-44b6-8e20-535e3b123890" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 148.198s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.156752] env[63202]: DEBUG oslo_concurrency.lockutils [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.901s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.157661] env[63202]: ERROR nova.compute.manager [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ad906a1e-35dc-4bb7-85b6-2672972103d0, please check neutron logs for more information. [ 793.157661] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] Traceback (most recent call last): [ 793.157661] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 793.157661] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] self.driver.spawn(context, instance, image_meta, [ 793.157661] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 793.157661] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] self._vmops.spawn(context, instance, image_meta, injected_files, [ 793.157661] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 793.157661] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] vm_ref = self.build_virtual_machine(instance, [ 793.157661] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 793.157661] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] vif_infos = vmwarevif.get_vif_info(self._session, [ 793.157661] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 793.157996] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] for vif in network_info: [ 793.157996] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 793.157996] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] return self._sync_wrapper(fn, *args, **kwargs) [ 793.157996] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 793.157996] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] self.wait() [ 793.157996] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 793.157996] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] self[:] = self._gt.wait() [ 793.157996] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 793.157996] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] return self._exit_event.wait() [ 793.157996] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 793.157996] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] result = hub.switch() [ 793.157996] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 793.157996] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] return self.greenlet.switch() [ 793.158385] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 793.158385] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] result = function(*args, **kwargs) [ 793.158385] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 793.158385] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] return func(*args, **kwargs) [ 793.158385] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 793.158385] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] raise e [ 793.158385] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 793.158385] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] nwinfo = self.network_api.allocate_for_instance( [ 793.158385] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 793.158385] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] created_port_ids = self._update_ports_for_instance( [ 793.158385] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 793.158385] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] with excutils.save_and_reraise_exception(): [ 793.158385] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 793.158833] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] self.force_reraise() [ 793.158833] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 793.158833] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] raise self.value [ 793.158833] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 793.158833] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] updated_port = self._update_port( [ 793.158833] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 793.158833] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] _ensure_no_port_binding_failure(port) [ 793.158833] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 793.158833] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] raise exception.PortBindingFailed(port_id=port['id']) [ 793.158833] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] nova.exception.PortBindingFailed: Binding failed for port ad906a1e-35dc-4bb7-85b6-2672972103d0, please check neutron logs for more information. [ 793.158833] env[63202]: ERROR nova.compute.manager [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] [ 793.159331] env[63202]: DEBUG nova.compute.utils [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] Binding failed for port ad906a1e-35dc-4bb7-85b6-2672972103d0, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 793.160221] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 13.138s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.162999] env[63202]: DEBUG nova.compute.manager [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] Build of instance e3a6ad78-4f46-42d4-935c-3cf310123530 was re-scheduled: Binding failed for port ad906a1e-35dc-4bb7-85b6-2672972103d0, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 793.163484] env[63202]: DEBUG nova.compute.manager [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 793.163734] env[63202]: DEBUG oslo_concurrency.lockutils [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] Acquiring lock "refresh_cache-e3a6ad78-4f46-42d4-935c-3cf310123530" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.163847] env[63202]: DEBUG oslo_concurrency.lockutils [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] Acquired lock "refresh_cache-e3a6ad78-4f46-42d4-935c-3cf310123530" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.164042] env[63202]: DEBUG nova.network.neutron [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 793.473969] env[63202]: INFO nova.compute.manager [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 8045881e-9bce-46e7-98c6-a7989f61a31e] Took 1.04 seconds to deallocate network for instance. [ 793.607282] env[63202]: DEBUG nova.compute.manager [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 793.686358] env[63202]: DEBUG nova.network.neutron [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 793.762924] env[63202]: DEBUG nova.network.neutron [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.797632] env[63202]: DEBUG oslo_concurrency.lockutils [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Acquiring lock "5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.797632] env[63202]: DEBUG oslo_concurrency.lockutils [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Lock "5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.129471] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.265583] env[63202]: DEBUG oslo_concurrency.lockutils [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] Releasing lock "refresh_cache-e3a6ad78-4f46-42d4-935c-3cf310123530" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.265721] env[63202]: DEBUG nova.compute.manager [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 794.265806] env[63202]: DEBUG nova.compute.manager [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 794.265969] env[63202]: DEBUG nova.network.neutron [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 794.285775] env[63202]: DEBUG nova.network.neutron [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 794.504878] env[63202]: INFO nova.scheduler.client.report [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Deleted allocations for instance 8045881e-9bce-46e7-98c6-a7989f61a31e [ 794.695358] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 8045881e-9bce-46e7-98c6-a7989f61a31e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 794.789396] env[63202]: DEBUG nova.network.neutron [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.014413] env[63202]: DEBUG oslo_concurrency.lockutils [None req-9a98b66b-052e-4a22-9bc6-0381c46263bc tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "8045881e-9bce-46e7-98c6-a7989f61a31e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 148.232s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.197949] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance e3a6ad78-4f46-42d4-935c-3cf310123530 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 795.198135] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance c9e0ee31-de9e-420d-8dad-380391d6f4e3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 795.291217] env[63202]: INFO nova.compute.manager [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] [instance: e3a6ad78-4f46-42d4-935c-3cf310123530] Took 1.03 seconds to deallocate network for instance. [ 795.519484] env[63202]: DEBUG nova.compute.manager [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 795.701030] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance e24b7994-3786-4709-bcb5-e0b2c9731f2a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 796.045089] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.204060] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 796.327035] env[63202]: INFO nova.scheduler.client.report [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] Deleted allocations for instance e3a6ad78-4f46-42d4-935c-3cf310123530 [ 796.709793] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 95d2a2d7-443b-43d6-97ff-fe36fd11002f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 796.834131] env[63202]: DEBUG oslo_concurrency.lockutils [None req-01c717be-075e-4910-8d8f-f5ca8016a429 tempest-ServerExternalEventsTest-1409432583 tempest-ServerExternalEventsTest-1409432583-project-member] Lock "e3a6ad78-4f46-42d4-935c-3cf310123530" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 145.743s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.213586] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 39d9b91b-c009-40e1-ad2b-ebef650188a4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 797.340939] env[63202]: DEBUG nova.compute.manager [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 797.347061] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "50854a4e-8e4e-431b-8220-01bf8906b1c0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.347061] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "50854a4e-8e4e-431b-8220-01bf8906b1c0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.716598] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 06bd8147-5479-4c5e-8ba2-6d3a205ba05e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 797.870395] env[63202]: DEBUG oslo_concurrency.lockutils [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.221428] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 798.724446] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 799.227918] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance a47dbb3e-62be-426d-acd7-bedc0faf65c5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 799.737440] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 574ddcba-db96-4bd1-bb9e-3f8846c13430 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 800.242040] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 57f446ee-edf3-40e7-96bf-8fdce62bb19f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 800.745760] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance ac48e21a-72ec-4c9d-8262-5d6cb36ec531 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 801.253657] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 69c66a35-3789-4c2f-b7cf-5f4136d53b94 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 801.756401] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 5d9f1803-e3f8-42a5-b6b7-2642463a079e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 802.260026] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance d674db7a-8cf0-46f8-b676-fced0767c105 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 802.762640] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance cedc3a06-2123-4c5e-a6c3-599a3efc3c65 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 803.050476] env[63202]: DEBUG oslo_concurrency.lockutils [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Acquiring lock "2243bb69-0dc5-49cd-b94e-73e703cbadc0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.050734] env[63202]: DEBUG oslo_concurrency.lockutils [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Lock "2243bb69-0dc5-49cd-b94e-73e703cbadc0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.265845] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 803.768589] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance a800fa5d-46a1-4a96-8ea2-26e34b869aa6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 804.271888] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 804.272250] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=63202) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 804.272466] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=63202) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 804.528056] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae96edb3-3f07-459c-b757-03ef1314ec9c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.535630] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bc51e74-87d7-4122-b398-2bf367909572 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.566222] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35242205-a844-45f9-b612-82148dbdb692 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.574122] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17fd6069-0f1f-4f86-bb63-67ecf5a88ecf {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.588070] env[63202]: DEBUG nova.compute.provider_tree [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 805.090878] env[63202]: DEBUG nova.scheduler.client.report [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 805.596618] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63202) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 805.596917] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 12.437s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.597275] env[63202]: DEBUG oslo_concurrency.lockutils [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 25.139s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.313636] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5523274-f1e1-486d-8208-4743aa2e5d32 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.321309] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bc2aed5-cadf-4032-ba5b-24101321aae2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.351042] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55935a3e-8adf-4070-b72b-53588d944693 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.358101] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f83f574b-b97e-4595-b662-753c8ba30b9c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.370802] env[63202]: DEBUG nova.compute.provider_tree [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 806.792059] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 806.875039] env[63202]: DEBUG nova.scheduler.client.report [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 807.297645] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 807.297805] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Starting heal instance info cache {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 807.297920] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Rebuilding the list of instances to heal {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 807.379265] env[63202]: DEBUG oslo_concurrency.lockutils [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.782s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.379888] env[63202]: ERROR nova.compute.manager [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f6092551-11d9-42f1-bd5f-f010c6f57650, please check neutron logs for more information. [ 807.379888] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] Traceback (most recent call last): [ 807.379888] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 807.379888] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] self.driver.spawn(context, instance, image_meta, [ 807.379888] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 807.379888] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 807.379888] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 807.379888] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] vm_ref = self.build_virtual_machine(instance, [ 807.379888] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 807.379888] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] vif_infos = vmwarevif.get_vif_info(self._session, [ 807.379888] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 807.380373] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] for vif in network_info: [ 807.380373] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 807.380373] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] return self._sync_wrapper(fn, *args, **kwargs) [ 807.380373] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 807.380373] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] self.wait() [ 807.380373] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 807.380373] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] self[:] = self._gt.wait() [ 807.380373] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 807.380373] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] return self._exit_event.wait() [ 807.380373] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 807.380373] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] current.throw(*self._exc) [ 807.380373] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 807.380373] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] result = function(*args, **kwargs) [ 807.380735] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 807.380735] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] return func(*args, **kwargs) [ 807.380735] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 807.380735] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] raise e [ 807.380735] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 807.380735] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] nwinfo = self.network_api.allocate_for_instance( [ 807.380735] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 807.380735] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] created_port_ids = self._update_ports_for_instance( [ 807.380735] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 807.380735] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] with excutils.save_and_reraise_exception(): [ 807.380735] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 807.380735] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] self.force_reraise() [ 807.380735] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 807.381088] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] raise self.value [ 807.381088] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 807.381088] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] updated_port = self._update_port( [ 807.381088] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 807.381088] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] _ensure_no_port_binding_failure(port) [ 807.381088] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 807.381088] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] raise exception.PortBindingFailed(port_id=port['id']) [ 807.381088] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] nova.exception.PortBindingFailed: Binding failed for port f6092551-11d9-42f1-bd5f-f010c6f57650, please check neutron logs for more information. [ 807.381088] env[63202]: ERROR nova.compute.manager [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] [ 807.381088] env[63202]: DEBUG nova.compute.utils [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] Binding failed for port f6092551-11d9-42f1-bd5f-f010c6f57650, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 807.383035] env[63202]: DEBUG nova.compute.manager [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] Build of instance c9e0ee31-de9e-420d-8dad-380391d6f4e3 was re-scheduled: Binding failed for port f6092551-11d9-42f1-bd5f-f010c6f57650, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 807.383035] env[63202]: DEBUG nova.compute.manager [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 807.383293] env[63202]: DEBUG oslo_concurrency.lockutils [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Acquiring lock "refresh_cache-c9e0ee31-de9e-420d-8dad-380391d6f4e3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.383293] env[63202]: DEBUG oslo_concurrency.lockutils [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Acquired lock "refresh_cache-c9e0ee31-de9e-420d-8dad-380391d6f4e3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.383367] env[63202]: DEBUG nova.network.neutron [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 807.384394] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.543s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.385816] env[63202]: INFO nova.compute.claims [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 807.802609] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Didn't find any instances for network info cache update. {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 807.907257] env[63202]: DEBUG nova.network.neutron [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 808.006113] env[63202]: DEBUG nova.network.neutron [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.510038] env[63202]: DEBUG oslo_concurrency.lockutils [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Releasing lock "refresh_cache-c9e0ee31-de9e-420d-8dad-380391d6f4e3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.510038] env[63202]: DEBUG nova.compute.manager [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 808.510038] env[63202]: DEBUG nova.compute.manager [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 808.510038] env[63202]: DEBUG nova.network.neutron [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 808.534838] env[63202]: DEBUG nova.network.neutron [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 808.617588] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c21f756b-2bf2-4222-8e95-e8db9116ed5d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.624786] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bacc67be-5980-42a6-af38-004aeb37bf3a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.653328] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a5d3f66-25f1-4aed-99b5-a8ae63efa94e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.659718] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef1fe66c-ef5c-406c-81ac-ec24d00ecd62 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.673502] env[63202]: DEBUG nova.compute.provider_tree [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 809.037670] env[63202]: DEBUG nova.network.neutron [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.176456] env[63202]: DEBUG nova.scheduler.client.report [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 809.539937] env[63202]: INFO nova.compute.manager [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: c9e0ee31-de9e-420d-8dad-380391d6f4e3] Took 1.03 seconds to deallocate network for instance. [ 809.681347] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.297s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.681837] env[63202]: DEBUG nova.compute.manager [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 809.685036] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.029s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.686566] env[63202]: INFO nova.compute.claims [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 810.186465] env[63202]: DEBUG nova.compute.utils [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 810.187901] env[63202]: DEBUG nova.compute.manager [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 810.188598] env[63202]: DEBUG nova.network.neutron [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 810.231937] env[63202]: DEBUG nova.policy [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2d26ff71671948518d7ae11efd247a4f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1228ad38d4f9419aa0699bbeeba5c0b7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 810.523980] env[63202]: DEBUG nova.network.neutron [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] Successfully created port: ccc1a7cc-c159-499e-9bf1-76d48a9127b3 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 810.567855] env[63202]: INFO nova.scheduler.client.report [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Deleted allocations for instance c9e0ee31-de9e-420d-8dad-380391d6f4e3 [ 810.691658] env[63202]: DEBUG nova.compute.manager [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 810.943992] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40d93ddc-3a9f-4920-895d-b861139f24b8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.952044] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5110aeb-72ae-4de4-ae50-1108122a66ae {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.984680] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08f209b8-77f3-4657-9d7a-3f24a6e91316 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.992399] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34cdc5dd-3d84-4e73-a7c1-0c642ca5f114 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.006655] env[63202]: DEBUG nova.compute.provider_tree [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 811.076479] env[63202]: DEBUG oslo_concurrency.lockutils [None req-86f4b214-848b-40a3-851f-e56fec84875f tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Lock "c9e0ee31-de9e-420d-8dad-380391d6f4e3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 158.863s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.470550] env[63202]: DEBUG nova.compute.manager [req-44e6c2fd-f6ff-49c8-9266-469ed02ef5dd req-462c41ad-e2f6-4841-ac27-142c41c4f9d1 service nova] [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] Received event network-changed-ccc1a7cc-c159-499e-9bf1-76d48a9127b3 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 811.470752] env[63202]: DEBUG nova.compute.manager [req-44e6c2fd-f6ff-49c8-9266-469ed02ef5dd req-462c41ad-e2f6-4841-ac27-142c41c4f9d1 service nova] [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] Refreshing instance network info cache due to event network-changed-ccc1a7cc-c159-499e-9bf1-76d48a9127b3. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 811.470960] env[63202]: DEBUG oslo_concurrency.lockutils [req-44e6c2fd-f6ff-49c8-9266-469ed02ef5dd req-462c41ad-e2f6-4841-ac27-142c41c4f9d1 service nova] Acquiring lock "refresh_cache-e24b7994-3786-4709-bcb5-e0b2c9731f2a" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.471227] env[63202]: DEBUG oslo_concurrency.lockutils [req-44e6c2fd-f6ff-49c8-9266-469ed02ef5dd req-462c41ad-e2f6-4841-ac27-142c41c4f9d1 service nova] Acquired lock "refresh_cache-e24b7994-3786-4709-bcb5-e0b2c9731f2a" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.471438] env[63202]: DEBUG nova.network.neutron [req-44e6c2fd-f6ff-49c8-9266-469ed02ef5dd req-462c41ad-e2f6-4841-ac27-142c41c4f9d1 service nova] [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] Refreshing network info cache for port ccc1a7cc-c159-499e-9bf1-76d48a9127b3 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 811.511081] env[63202]: DEBUG nova.scheduler.client.report [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 811.579297] env[63202]: DEBUG nova.compute.manager [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 811.681870] env[63202]: ERROR nova.compute.manager [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ccc1a7cc-c159-499e-9bf1-76d48a9127b3, please check neutron logs for more information. [ 811.681870] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 811.681870] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 811.681870] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 811.681870] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 811.681870] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 811.681870] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 811.681870] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 811.681870] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 811.681870] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 811.681870] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 811.681870] env[63202]: ERROR nova.compute.manager raise self.value [ 811.681870] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 811.681870] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 811.681870] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 811.681870] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 811.682438] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 811.682438] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 811.682438] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ccc1a7cc-c159-499e-9bf1-76d48a9127b3, please check neutron logs for more information. [ 811.682438] env[63202]: ERROR nova.compute.manager [ 811.682438] env[63202]: Traceback (most recent call last): [ 811.682438] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 811.682438] env[63202]: listener.cb(fileno) [ 811.682438] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 811.682438] env[63202]: result = function(*args, **kwargs) [ 811.682438] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 811.682438] env[63202]: return func(*args, **kwargs) [ 811.682438] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 811.682438] env[63202]: raise e [ 811.682438] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 811.682438] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 811.682438] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 811.682438] env[63202]: created_port_ids = self._update_ports_for_instance( [ 811.682438] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 811.682438] env[63202]: with excutils.save_and_reraise_exception(): [ 811.682438] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 811.682438] env[63202]: self.force_reraise() [ 811.682438] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 811.682438] env[63202]: raise self.value [ 811.682438] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 811.682438] env[63202]: updated_port = self._update_port( [ 811.682438] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 811.682438] env[63202]: _ensure_no_port_binding_failure(port) [ 811.682438] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 811.682438] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 811.683370] env[63202]: nova.exception.PortBindingFailed: Binding failed for port ccc1a7cc-c159-499e-9bf1-76d48a9127b3, please check neutron logs for more information. [ 811.683370] env[63202]: Removing descriptor: 16 [ 811.704442] env[63202]: DEBUG nova.compute.manager [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 811.729858] env[63202]: DEBUG nova.virt.hardware [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 811.730178] env[63202]: DEBUG nova.virt.hardware [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 811.730385] env[63202]: DEBUG nova.virt.hardware [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 811.730608] env[63202]: DEBUG nova.virt.hardware [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 811.730760] env[63202]: DEBUG nova.virt.hardware [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 811.730905] env[63202]: DEBUG nova.virt.hardware [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 811.731122] env[63202]: DEBUG nova.virt.hardware [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 811.731280] env[63202]: DEBUG nova.virt.hardware [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 811.731439] env[63202]: DEBUG nova.virt.hardware [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 811.731594] env[63202]: DEBUG nova.virt.hardware [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 811.731759] env[63202]: DEBUG nova.virt.hardware [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 811.732907] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60f27307-74fa-4b74-b1a0-faaaf010349f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.740429] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4290adea-9253-41d2-bc09-dc779ae1e996 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.753584] env[63202]: ERROR nova.compute.manager [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ccc1a7cc-c159-499e-9bf1-76d48a9127b3, please check neutron logs for more information. [ 811.753584] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] Traceback (most recent call last): [ 811.753584] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 811.753584] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] yield resources [ 811.753584] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 811.753584] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] self.driver.spawn(context, instance, image_meta, [ 811.753584] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 811.753584] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 811.753584] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 811.753584] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] vm_ref = self.build_virtual_machine(instance, [ 811.753584] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 811.753936] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] vif_infos = vmwarevif.get_vif_info(self._session, [ 811.753936] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 811.753936] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] for vif in network_info: [ 811.753936] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 811.753936] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] return self._sync_wrapper(fn, *args, **kwargs) [ 811.753936] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 811.753936] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] self.wait() [ 811.753936] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 811.753936] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] self[:] = self._gt.wait() [ 811.753936] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 811.753936] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] return self._exit_event.wait() [ 811.753936] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 811.753936] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] current.throw(*self._exc) [ 811.754385] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 811.754385] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] result = function(*args, **kwargs) [ 811.754385] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 811.754385] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] return func(*args, **kwargs) [ 811.754385] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 811.754385] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] raise e [ 811.754385] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 811.754385] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] nwinfo = self.network_api.allocate_for_instance( [ 811.754385] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 811.754385] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] created_port_ids = self._update_ports_for_instance( [ 811.754385] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 811.754385] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] with excutils.save_and_reraise_exception(): [ 811.754385] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 811.754809] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] self.force_reraise() [ 811.754809] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 811.754809] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] raise self.value [ 811.754809] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 811.754809] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] updated_port = self._update_port( [ 811.754809] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 811.754809] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] _ensure_no_port_binding_failure(port) [ 811.754809] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 811.754809] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] raise exception.PortBindingFailed(port_id=port['id']) [ 811.754809] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] nova.exception.PortBindingFailed: Binding failed for port ccc1a7cc-c159-499e-9bf1-76d48a9127b3, please check neutron logs for more information. [ 811.754809] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] [ 811.754809] env[63202]: INFO nova.compute.manager [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] Terminating instance [ 811.756152] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Acquiring lock "refresh_cache-e24b7994-3786-4709-bcb5-e0b2c9731f2a" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.996200] env[63202]: DEBUG nova.network.neutron [req-44e6c2fd-f6ff-49c8-9266-469ed02ef5dd req-462c41ad-e2f6-4841-ac27-142c41c4f9d1 service nova] [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 812.015088] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.330s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.015604] env[63202]: DEBUG nova.compute.manager [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 812.018132] env[63202]: DEBUG oslo_concurrency.lockutils [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.493s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.019540] env[63202]: INFO nova.compute.claims [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 812.103820] env[63202]: DEBUG nova.network.neutron [req-44e6c2fd-f6ff-49c8-9266-469ed02ef5dd req-462c41ad-e2f6-4841-ac27-142c41c4f9d1 service nova] [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.105624] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.523710] env[63202]: DEBUG nova.compute.utils [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 812.526780] env[63202]: DEBUG nova.compute.manager [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 812.526945] env[63202]: DEBUG nova.network.neutron [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 812.572922] env[63202]: DEBUG nova.policy [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '91c024b1ffe34a778a7a86292ebd64cb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '98c04bf0db3147b896627a3da19d40a8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 812.609107] env[63202]: DEBUG oslo_concurrency.lockutils [req-44e6c2fd-f6ff-49c8-9266-469ed02ef5dd req-462c41ad-e2f6-4841-ac27-142c41c4f9d1 service nova] Releasing lock "refresh_cache-e24b7994-3786-4709-bcb5-e0b2c9731f2a" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.609598] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Acquired lock "refresh_cache-e24b7994-3786-4709-bcb5-e0b2c9731f2a" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.609786] env[63202]: DEBUG nova.network.neutron [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 813.029751] env[63202]: DEBUG nova.compute.manager [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 813.038346] env[63202]: DEBUG nova.network.neutron [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] Successfully created port: 965fe5c5-bcdd-42fb-912d-ad9870e37bf0 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 813.137601] env[63202]: DEBUG nova.network.neutron [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 813.221800] env[63202]: DEBUG nova.network.neutron [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.310368] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1880ece9-219e-433d-a805-346f715166b4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.318427] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e63dc48-2710-4f5e-a5bd-e5a8570952d0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.347685] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89608086-2d16-4f5b-a4df-d17b7161033b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.355023] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d693d3b-1c07-4fb9-b5b6-c6d786ff7511 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.369283] env[63202]: DEBUG nova.compute.provider_tree [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 813.501306] env[63202]: DEBUG nova.compute.manager [req-cb9d1cad-de0d-41ee-80c9-19f4a2a09988 req-f1983810-1982-4cb2-9894-a70722331459 service nova] [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] Received event network-vif-deleted-ccc1a7cc-c159-499e-9bf1-76d48a9127b3 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 813.724823] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Releasing lock "refresh_cache-e24b7994-3786-4709-bcb5-e0b2c9731f2a" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.725274] env[63202]: DEBUG nova.compute.manager [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 813.725464] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 813.725784] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-78084eeb-1aaa-4555-94a2-688968fd4064 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.735527] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f91126a-4983-4f51-9da1-a74f529cecc0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.757842] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e24b7994-3786-4709-bcb5-e0b2c9731f2a could not be found. [ 813.758053] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 813.758236] env[63202]: INFO nova.compute.manager [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] Took 0.03 seconds to destroy the instance on the hypervisor. [ 813.758460] env[63202]: DEBUG oslo.service.loopingcall [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 813.758707] env[63202]: DEBUG nova.compute.manager [-] [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 813.758760] env[63202]: DEBUG nova.network.neutron [-] [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 813.775955] env[63202]: DEBUG nova.network.neutron [-] [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 813.872612] env[63202]: DEBUG nova.scheduler.client.report [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 814.041703] env[63202]: DEBUG nova.compute.manager [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 814.068013] env[63202]: DEBUG nova.virt.hardware [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 814.068013] env[63202]: DEBUG nova.virt.hardware [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 814.068013] env[63202]: DEBUG nova.virt.hardware [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 814.068180] env[63202]: DEBUG nova.virt.hardware [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 814.068219] env[63202]: DEBUG nova.virt.hardware [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 814.068358] env[63202]: DEBUG nova.virt.hardware [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 814.069989] env[63202]: DEBUG nova.virt.hardware [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 814.069989] env[63202]: DEBUG nova.virt.hardware [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 814.069989] env[63202]: DEBUG nova.virt.hardware [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 814.069989] env[63202]: DEBUG nova.virt.hardware [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 814.069989] env[63202]: DEBUG nova.virt.hardware [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 814.070304] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eb44959-c555-46b6-82d8-1a132cb1903a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.073189] env[63202]: ERROR nova.compute.manager [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 965fe5c5-bcdd-42fb-912d-ad9870e37bf0, please check neutron logs for more information. [ 814.073189] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 814.073189] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 814.073189] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 814.073189] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 814.073189] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 814.073189] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 814.073189] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 814.073189] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 814.073189] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 814.073189] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 814.073189] env[63202]: ERROR nova.compute.manager raise self.value [ 814.073189] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 814.073189] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 814.073189] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 814.073189] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 814.073674] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 814.073674] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 814.073674] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 965fe5c5-bcdd-42fb-912d-ad9870e37bf0, please check neutron logs for more information. [ 814.073674] env[63202]: ERROR nova.compute.manager [ 814.073674] env[63202]: Traceback (most recent call last): [ 814.073674] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 814.073674] env[63202]: listener.cb(fileno) [ 814.073674] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 814.073674] env[63202]: result = function(*args, **kwargs) [ 814.073674] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 814.073674] env[63202]: return func(*args, **kwargs) [ 814.073674] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 814.073674] env[63202]: raise e [ 814.073674] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 814.073674] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 814.073674] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 814.073674] env[63202]: created_port_ids = self._update_ports_for_instance( [ 814.073674] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 814.073674] env[63202]: with excutils.save_and_reraise_exception(): [ 814.073674] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 814.073674] env[63202]: self.force_reraise() [ 814.073674] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 814.073674] env[63202]: raise self.value [ 814.073674] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 814.073674] env[63202]: updated_port = self._update_port( [ 814.073674] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 814.073674] env[63202]: _ensure_no_port_binding_failure(port) [ 814.073674] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 814.073674] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 814.074586] env[63202]: nova.exception.PortBindingFailed: Binding failed for port 965fe5c5-bcdd-42fb-912d-ad9870e37bf0, please check neutron logs for more information. [ 814.074586] env[63202]: Removing descriptor: 16 [ 814.078892] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4d508e2-a3a1-4902-b374-405edf38f513 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.092525] env[63202]: ERROR nova.compute.manager [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 965fe5c5-bcdd-42fb-912d-ad9870e37bf0, please check neutron logs for more information. [ 814.092525] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] Traceback (most recent call last): [ 814.092525] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 814.092525] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] yield resources [ 814.092525] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 814.092525] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] self.driver.spawn(context, instance, image_meta, [ 814.092525] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 814.092525] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 814.092525] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 814.092525] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] vm_ref = self.build_virtual_machine(instance, [ 814.092525] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 814.092867] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] vif_infos = vmwarevif.get_vif_info(self._session, [ 814.092867] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 814.092867] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] for vif in network_info: [ 814.092867] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 814.092867] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] return self._sync_wrapper(fn, *args, **kwargs) [ 814.092867] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 814.092867] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] self.wait() [ 814.092867] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 814.092867] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] self[:] = self._gt.wait() [ 814.092867] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 814.092867] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] return self._exit_event.wait() [ 814.092867] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 814.092867] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] current.throw(*self._exc) [ 814.093368] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 814.093368] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] result = function(*args, **kwargs) [ 814.093368] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 814.093368] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] return func(*args, **kwargs) [ 814.093368] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 814.093368] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] raise e [ 814.093368] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 814.093368] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] nwinfo = self.network_api.allocate_for_instance( [ 814.093368] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 814.093368] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] created_port_ids = self._update_ports_for_instance( [ 814.093368] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 814.093368] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] with excutils.save_and_reraise_exception(): [ 814.093368] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 814.093854] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] self.force_reraise() [ 814.093854] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 814.093854] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] raise self.value [ 814.093854] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 814.093854] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] updated_port = self._update_port( [ 814.093854] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 814.093854] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] _ensure_no_port_binding_failure(port) [ 814.093854] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 814.093854] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] raise exception.PortBindingFailed(port_id=port['id']) [ 814.093854] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] nova.exception.PortBindingFailed: Binding failed for port 965fe5c5-bcdd-42fb-912d-ad9870e37bf0, please check neutron logs for more information. [ 814.093854] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] [ 814.093854] env[63202]: INFO nova.compute.manager [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] Terminating instance [ 814.094838] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] Acquiring lock "refresh_cache-d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.095102] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] Acquired lock "refresh_cache-d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.095163] env[63202]: DEBUG nova.network.neutron [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 814.278398] env[63202]: DEBUG nova.network.neutron [-] [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 814.382053] env[63202]: DEBUG oslo_concurrency.lockutils [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.363s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.382053] env[63202]: DEBUG nova.compute.manager [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 814.384724] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.935s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.386511] env[63202]: INFO nova.compute.claims [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 814.611274] env[63202]: DEBUG nova.network.neutron [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 814.694485] env[63202]: DEBUG nova.network.neutron [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 814.781039] env[63202]: INFO nova.compute.manager [-] [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] Took 1.02 seconds to deallocate network for instance. [ 814.783202] env[63202]: DEBUG nova.compute.claims [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 814.783417] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.890183] env[63202]: DEBUG nova.compute.utils [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 814.893462] env[63202]: DEBUG nova.compute.manager [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 814.893663] env[63202]: DEBUG nova.network.neutron [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 814.930378] env[63202]: DEBUG nova.policy [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '087616024ab9407a9dee7ea4d109ad94', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a231702f313a4ffe81c67f8f0ce83058', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 815.196672] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] Releasing lock "refresh_cache-d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.197112] env[63202]: DEBUG nova.compute.manager [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 815.197376] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 815.197682] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9482b66d-fb88-4022-b58f-bef4c8a70b9d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.207012] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-964aa736-1405-48e1-a2e0-f7f829ae1d41 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.229083] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6 could not be found. [ 815.229360] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 815.229547] env[63202]: INFO nova.compute.manager [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] Took 0.03 seconds to destroy the instance on the hypervisor. [ 815.229786] env[63202]: DEBUG oslo.service.loopingcall [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 815.230070] env[63202]: DEBUG nova.compute.manager [-] [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 815.230120] env[63202]: DEBUG nova.network.neutron [-] [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 815.245365] env[63202]: DEBUG nova.network.neutron [-] [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 815.353922] env[63202]: DEBUG nova.network.neutron [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] Successfully created port: 560f6e99-7a69-440c-bee5-e88a343edd85 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 815.394065] env[63202]: DEBUG nova.compute.manager [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 815.581566] env[63202]: DEBUG nova.compute.manager [req-e4bb76fb-1264-4e8c-9fcd-bb8610bc05ad req-e51d4ea9-340c-42dd-a70e-9fc9db33bcb3 service nova] [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] Received event network-changed-965fe5c5-bcdd-42fb-912d-ad9870e37bf0 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 815.582129] env[63202]: DEBUG nova.compute.manager [req-e4bb76fb-1264-4e8c-9fcd-bb8610bc05ad req-e51d4ea9-340c-42dd-a70e-9fc9db33bcb3 service nova] [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] Refreshing instance network info cache due to event network-changed-965fe5c5-bcdd-42fb-912d-ad9870e37bf0. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 815.582129] env[63202]: DEBUG oslo_concurrency.lockutils [req-e4bb76fb-1264-4e8c-9fcd-bb8610bc05ad req-e51d4ea9-340c-42dd-a70e-9fc9db33bcb3 service nova] Acquiring lock "refresh_cache-d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.582220] env[63202]: DEBUG oslo_concurrency.lockutils [req-e4bb76fb-1264-4e8c-9fcd-bb8610bc05ad req-e51d4ea9-340c-42dd-a70e-9fc9db33bcb3 service nova] Acquired lock "refresh_cache-d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.582413] env[63202]: DEBUG nova.network.neutron [req-e4bb76fb-1264-4e8c-9fcd-bb8610bc05ad req-e51d4ea9-340c-42dd-a70e-9fc9db33bcb3 service nova] [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] Refreshing network info cache for port 965fe5c5-bcdd-42fb-912d-ad9870e37bf0 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 815.674697] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fbbcb95-767a-413e-b697-d63569a13136 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.682694] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f72884a7-b468-4198-9274-7b94e08929ff {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.413083] env[63202]: DEBUG nova.network.neutron [-] [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.418338] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73754b91-ade2-4f2f-8b17-87029079ba1d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.428154] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b3485c8-4ab8-49a8-a61e-2a55f75decec {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.435417] env[63202]: DEBUG nova.network.neutron [req-e4bb76fb-1264-4e8c-9fcd-bb8610bc05ad req-e51d4ea9-340c-42dd-a70e-9fc9db33bcb3 service nova] [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 816.444840] env[63202]: DEBUG nova.compute.provider_tree [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 816.521710] env[63202]: DEBUG nova.network.neutron [req-e4bb76fb-1264-4e8c-9fcd-bb8610bc05ad req-e51d4ea9-340c-42dd-a70e-9fc9db33bcb3 service nova] [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.731491] env[63202]: ERROR nova.compute.manager [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 560f6e99-7a69-440c-bee5-e88a343edd85, please check neutron logs for more information. [ 816.731491] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 816.731491] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 816.731491] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 816.731491] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 816.731491] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 816.731491] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 816.731491] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 816.731491] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 816.731491] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 816.731491] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 816.731491] env[63202]: ERROR nova.compute.manager raise self.value [ 816.731491] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 816.731491] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 816.731491] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 816.731491] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 816.732050] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 816.732050] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 816.732050] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 560f6e99-7a69-440c-bee5-e88a343edd85, please check neutron logs for more information. [ 816.732050] env[63202]: ERROR nova.compute.manager [ 816.732050] env[63202]: Traceback (most recent call last): [ 816.732050] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 816.732050] env[63202]: listener.cb(fileno) [ 816.732050] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 816.732050] env[63202]: result = function(*args, **kwargs) [ 816.732050] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 816.732050] env[63202]: return func(*args, **kwargs) [ 816.732050] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 816.732050] env[63202]: raise e [ 816.732050] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 816.732050] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 816.732050] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 816.732050] env[63202]: created_port_ids = self._update_ports_for_instance( [ 816.732050] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 816.732050] env[63202]: with excutils.save_and_reraise_exception(): [ 816.732050] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 816.732050] env[63202]: self.force_reraise() [ 816.732050] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 816.732050] env[63202]: raise self.value [ 816.732050] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 816.732050] env[63202]: updated_port = self._update_port( [ 816.732050] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 816.732050] env[63202]: _ensure_no_port_binding_failure(port) [ 816.732050] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 816.732050] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 816.732969] env[63202]: nova.exception.PortBindingFailed: Binding failed for port 560f6e99-7a69-440c-bee5-e88a343edd85, please check neutron logs for more information. [ 816.732969] env[63202]: Removing descriptor: 16 [ 816.918049] env[63202]: DEBUG nova.compute.manager [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 816.920102] env[63202]: INFO nova.compute.manager [-] [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] Took 1.69 seconds to deallocate network for instance. [ 816.922632] env[63202]: DEBUG nova.compute.claims [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 816.922802] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.942261] env[63202]: DEBUG nova.virt.hardware [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 816.942477] env[63202]: DEBUG nova.virt.hardware [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 816.942627] env[63202]: DEBUG nova.virt.hardware [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 816.942798] env[63202]: DEBUG nova.virt.hardware [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 816.942941] env[63202]: DEBUG nova.virt.hardware [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 816.943091] env[63202]: DEBUG nova.virt.hardware [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 816.943291] env[63202]: DEBUG nova.virt.hardware [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 816.943447] env[63202]: DEBUG nova.virt.hardware [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 816.943605] env[63202]: DEBUG nova.virt.hardware [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 816.943761] env[63202]: DEBUG nova.virt.hardware [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 816.943923] env[63202]: DEBUG nova.virt.hardware [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 816.944752] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79e368dc-70e6-4300-b3c7-3570c1a5839e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.947643] env[63202]: DEBUG nova.scheduler.client.report [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 816.955669] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e134bffb-6ffb-42bf-9d1a-60ee1a2cad3a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.969364] env[63202]: ERROR nova.compute.manager [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 560f6e99-7a69-440c-bee5-e88a343edd85, please check neutron logs for more information. [ 816.969364] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] Traceback (most recent call last): [ 816.969364] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 816.969364] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] yield resources [ 816.969364] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 816.969364] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] self.driver.spawn(context, instance, image_meta, [ 816.969364] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 816.969364] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 816.969364] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 816.969364] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] vm_ref = self.build_virtual_machine(instance, [ 816.969364] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 816.969714] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] vif_infos = vmwarevif.get_vif_info(self._session, [ 816.969714] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 816.969714] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] for vif in network_info: [ 816.969714] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 816.969714] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] return self._sync_wrapper(fn, *args, **kwargs) [ 816.969714] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 816.969714] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] self.wait() [ 816.969714] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 816.969714] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] self[:] = self._gt.wait() [ 816.969714] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 816.969714] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] return self._exit_event.wait() [ 816.969714] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 816.969714] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] current.throw(*self._exc) [ 816.970265] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 816.970265] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] result = function(*args, **kwargs) [ 816.970265] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 816.970265] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] return func(*args, **kwargs) [ 816.970265] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 816.970265] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] raise e [ 816.970265] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 816.970265] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] nwinfo = self.network_api.allocate_for_instance( [ 816.970265] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 816.970265] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] created_port_ids = self._update_ports_for_instance( [ 816.970265] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 816.970265] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] with excutils.save_and_reraise_exception(): [ 816.970265] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 816.970678] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] self.force_reraise() [ 816.970678] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 816.970678] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] raise self.value [ 816.970678] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 816.970678] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] updated_port = self._update_port( [ 816.970678] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 816.970678] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] _ensure_no_port_binding_failure(port) [ 816.970678] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 816.970678] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] raise exception.PortBindingFailed(port_id=port['id']) [ 816.970678] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] nova.exception.PortBindingFailed: Binding failed for port 560f6e99-7a69-440c-bee5-e88a343edd85, please check neutron logs for more information. [ 816.970678] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] [ 816.970678] env[63202]: INFO nova.compute.manager [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] Terminating instance [ 816.971760] env[63202]: DEBUG oslo_concurrency.lockutils [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] Acquiring lock "refresh_cache-95d2a2d7-443b-43d6-97ff-fe36fd11002f" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.971915] env[63202]: DEBUG oslo_concurrency.lockutils [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] Acquired lock "refresh_cache-95d2a2d7-443b-43d6-97ff-fe36fd11002f" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.972083] env[63202]: DEBUG nova.network.neutron [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 817.025032] env[63202]: DEBUG oslo_concurrency.lockutils [req-e4bb76fb-1264-4e8c-9fcd-bb8610bc05ad req-e51d4ea9-340c-42dd-a70e-9fc9db33bcb3 service nova] Releasing lock "refresh_cache-d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.025123] env[63202]: DEBUG nova.compute.manager [req-e4bb76fb-1264-4e8c-9fcd-bb8610bc05ad req-e51d4ea9-340c-42dd-a70e-9fc9db33bcb3 service nova] [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] Received event network-vif-deleted-965fe5c5-bcdd-42fb-912d-ad9870e37bf0 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 817.452590] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.068s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.453178] env[63202]: DEBUG nova.compute.manager [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 817.456062] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.067s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.458865] env[63202]: INFO nova.compute.claims [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 817.493817] env[63202]: DEBUG nova.network.neutron [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 817.591865] env[63202]: DEBUG nova.network.neutron [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.632309] env[63202]: DEBUG nova.compute.manager [req-4cb1cb64-7227-49fb-a975-4d705db8f1b1 req-b8db6d7c-0295-4f77-8017-abeec09be95c service nova] [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] Received event network-changed-560f6e99-7a69-440c-bee5-e88a343edd85 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 817.632309] env[63202]: DEBUG nova.compute.manager [req-4cb1cb64-7227-49fb-a975-4d705db8f1b1 req-b8db6d7c-0295-4f77-8017-abeec09be95c service nova] [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] Refreshing instance network info cache due to event network-changed-560f6e99-7a69-440c-bee5-e88a343edd85. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 817.632309] env[63202]: DEBUG oslo_concurrency.lockutils [req-4cb1cb64-7227-49fb-a975-4d705db8f1b1 req-b8db6d7c-0295-4f77-8017-abeec09be95c service nova] Acquiring lock "refresh_cache-95d2a2d7-443b-43d6-97ff-fe36fd11002f" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.964869] env[63202]: DEBUG nova.compute.utils [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 817.966271] env[63202]: DEBUG nova.compute.manager [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 817.966431] env[63202]: DEBUG nova.network.neutron [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 818.004338] env[63202]: DEBUG nova.policy [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2d292997e4c34051acc645b4a65dc369', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5e914f87450949b1a39866e8cfa3a0eb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 818.097721] env[63202]: DEBUG oslo_concurrency.lockutils [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] Releasing lock "refresh_cache-95d2a2d7-443b-43d6-97ff-fe36fd11002f" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.098224] env[63202]: DEBUG nova.compute.manager [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 818.099141] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 818.099568] env[63202]: DEBUG oslo_concurrency.lockutils [req-4cb1cb64-7227-49fb-a975-4d705db8f1b1 req-b8db6d7c-0295-4f77-8017-abeec09be95c service nova] Acquired lock "refresh_cache-95d2a2d7-443b-43d6-97ff-fe36fd11002f" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.099746] env[63202]: DEBUG nova.network.neutron [req-4cb1cb64-7227-49fb-a975-4d705db8f1b1 req-b8db6d7c-0295-4f77-8017-abeec09be95c service nova] [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] Refreshing network info cache for port 560f6e99-7a69-440c-bee5-e88a343edd85 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 818.100838] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0e93add5-1e08-48e6-b13a-370319da074d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.111132] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c6ac1c5-15d7-451a-a49f-7029e88abd9e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.132771] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 95d2a2d7-443b-43d6-97ff-fe36fd11002f could not be found. [ 818.132986] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 818.133176] env[63202]: INFO nova.compute.manager [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] Took 0.03 seconds to destroy the instance on the hypervisor. [ 818.133410] env[63202]: DEBUG oslo.service.loopingcall [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 818.133625] env[63202]: DEBUG nova.compute.manager [-] [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 818.133714] env[63202]: DEBUG nova.network.neutron [-] [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 818.148908] env[63202]: DEBUG nova.network.neutron [-] [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 818.253267] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 818.253504] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 818.253680] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Cleaning up deleted instances {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 818.312319] env[63202]: DEBUG nova.network.neutron [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] Successfully created port: a9945f05-55e7-4534-a38e-d921a091bb11 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 818.470637] env[63202]: DEBUG nova.compute.manager [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 818.619995] env[63202]: DEBUG nova.network.neutron [req-4cb1cb64-7227-49fb-a975-4d705db8f1b1 req-b8db6d7c-0295-4f77-8017-abeec09be95c service nova] [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 818.653983] env[63202]: DEBUG nova.network.neutron [-] [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.725239] env[63202]: DEBUG nova.network.neutron [req-4cb1cb64-7227-49fb-a975-4d705db8f1b1 req-b8db6d7c-0295-4f77-8017-abeec09be95c service nova] [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.758997] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] There are 5 instances to clean {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 818.759400] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: c8d27e19-0f3c-4cc3-826b-e07b3a635e5e] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 818.761918] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-463f33c4-4f94-432a-b435-8fb704e3be3b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.770199] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08e17439-996f-400a-b361-efa5aada0c00 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.802450] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb276bad-e24b-4063-8332-aba1c8953c41 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.811083] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21a5bda3-0731-4687-b883-f16dff3d0349 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.823070] env[63202]: DEBUG nova.compute.provider_tree [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 819.157392] env[63202]: INFO nova.compute.manager [-] [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] Took 1.02 seconds to deallocate network for instance. [ 819.160485] env[63202]: DEBUG nova.compute.claims [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 819.160485] env[63202]: DEBUG oslo_concurrency.lockutils [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.226798] env[63202]: DEBUG oslo_concurrency.lockutils [req-4cb1cb64-7227-49fb-a975-4d705db8f1b1 req-b8db6d7c-0295-4f77-8017-abeec09be95c service nova] Releasing lock "refresh_cache-95d2a2d7-443b-43d6-97ff-fe36fd11002f" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.227065] env[63202]: DEBUG nova.compute.manager [req-4cb1cb64-7227-49fb-a975-4d705db8f1b1 req-b8db6d7c-0295-4f77-8017-abeec09be95c service nova] [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] Received event network-vif-deleted-560f6e99-7a69-440c-bee5-e88a343edd85 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 819.265584] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: e9b604fb-1d88-425c-92b7-008ddaec7c1a] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 819.325761] env[63202]: DEBUG nova.scheduler.client.report [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 819.479213] env[63202]: DEBUG nova.compute.manager [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 819.507016] env[63202]: DEBUG nova.virt.hardware [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 819.507016] env[63202]: DEBUG nova.virt.hardware [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 819.507016] env[63202]: DEBUG nova.virt.hardware [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 819.507016] env[63202]: DEBUG nova.virt.hardware [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 819.507274] env[63202]: DEBUG nova.virt.hardware [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 819.507274] env[63202]: DEBUG nova.virt.hardware [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 819.507274] env[63202]: DEBUG nova.virt.hardware [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 819.507274] env[63202]: DEBUG nova.virt.hardware [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 819.507599] env[63202]: DEBUG nova.virt.hardware [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 819.507905] env[63202]: DEBUG nova.virt.hardware [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 819.508208] env[63202]: DEBUG nova.virt.hardware [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 819.509164] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fe89f12-7393-4b67-bb5b-9e3a73f48c07 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.517394] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddffe866-a19d-4d34-b647-810ca1ee2941 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.561676] env[63202]: ERROR nova.compute.manager [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a9945f05-55e7-4534-a38e-d921a091bb11, please check neutron logs for more information. [ 819.561676] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 819.561676] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 819.561676] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 819.561676] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 819.561676] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 819.561676] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 819.561676] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 819.561676] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 819.561676] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 819.561676] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 819.561676] env[63202]: ERROR nova.compute.manager raise self.value [ 819.561676] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 819.561676] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 819.561676] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 819.561676] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 819.562275] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 819.562275] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 819.562275] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a9945f05-55e7-4534-a38e-d921a091bb11, please check neutron logs for more information. [ 819.562275] env[63202]: ERROR nova.compute.manager [ 819.562275] env[63202]: Traceback (most recent call last): [ 819.562275] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 819.562275] env[63202]: listener.cb(fileno) [ 819.562275] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 819.562275] env[63202]: result = function(*args, **kwargs) [ 819.562275] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 819.562275] env[63202]: return func(*args, **kwargs) [ 819.562275] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 819.562275] env[63202]: raise e [ 819.562275] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 819.562275] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 819.562275] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 819.562275] env[63202]: created_port_ids = self._update_ports_for_instance( [ 819.562275] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 819.562275] env[63202]: with excutils.save_and_reraise_exception(): [ 819.562275] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 819.562275] env[63202]: self.force_reraise() [ 819.562275] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 819.562275] env[63202]: raise self.value [ 819.562275] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 819.562275] env[63202]: updated_port = self._update_port( [ 819.562275] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 819.562275] env[63202]: _ensure_no_port_binding_failure(port) [ 819.562275] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 819.562275] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 819.563207] env[63202]: nova.exception.PortBindingFailed: Binding failed for port a9945f05-55e7-4534-a38e-d921a091bb11, please check neutron logs for more information. [ 819.563207] env[63202]: Removing descriptor: 16 [ 819.563207] env[63202]: ERROR nova.compute.manager [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a9945f05-55e7-4534-a38e-d921a091bb11, please check neutron logs for more information. [ 819.563207] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] Traceback (most recent call last): [ 819.563207] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 819.563207] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] yield resources [ 819.563207] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 819.563207] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] self.driver.spawn(context, instance, image_meta, [ 819.563207] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 819.563207] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 819.563207] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 819.563207] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] vm_ref = self.build_virtual_machine(instance, [ 819.563817] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 819.563817] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] vif_infos = vmwarevif.get_vif_info(self._session, [ 819.563817] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 819.563817] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] for vif in network_info: [ 819.563817] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 819.563817] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] return self._sync_wrapper(fn, *args, **kwargs) [ 819.563817] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 819.563817] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] self.wait() [ 819.563817] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 819.563817] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] self[:] = self._gt.wait() [ 819.563817] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 819.563817] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] return self._exit_event.wait() [ 819.563817] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 819.564329] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] result = hub.switch() [ 819.564329] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 819.564329] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] return self.greenlet.switch() [ 819.564329] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 819.564329] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] result = function(*args, **kwargs) [ 819.564329] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 819.564329] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] return func(*args, **kwargs) [ 819.564329] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 819.564329] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] raise e [ 819.564329] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 819.564329] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] nwinfo = self.network_api.allocate_for_instance( [ 819.564329] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 819.564329] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] created_port_ids = self._update_ports_for_instance( [ 819.564740] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 819.564740] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] with excutils.save_and_reraise_exception(): [ 819.564740] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 819.564740] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] self.force_reraise() [ 819.564740] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 819.564740] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] raise self.value [ 819.564740] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 819.564740] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] updated_port = self._update_port( [ 819.564740] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 819.564740] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] _ensure_no_port_binding_failure(port) [ 819.564740] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 819.564740] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] raise exception.PortBindingFailed(port_id=port['id']) [ 819.565139] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] nova.exception.PortBindingFailed: Binding failed for port a9945f05-55e7-4534-a38e-d921a091bb11, please check neutron logs for more information. [ 819.565139] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] [ 819.565139] env[63202]: INFO nova.compute.manager [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] Terminating instance [ 819.565139] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquiring lock "refresh_cache-39d9b91b-c009-40e1-ad2b-ebef650188a4" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.565139] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquired lock "refresh_cache-39d9b91b-c009-40e1-ad2b-ebef650188a4" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.565139] env[63202]: DEBUG nova.network.neutron [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 819.655247] env[63202]: DEBUG nova.compute.manager [req-4709b7ef-80bb-4d16-80cb-a36b7167fe80 req-eb1b0105-2ed7-4407-8e5c-ff408368721b service nova] [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] Received event network-changed-a9945f05-55e7-4534-a38e-d921a091bb11 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 819.655461] env[63202]: DEBUG nova.compute.manager [req-4709b7ef-80bb-4d16-80cb-a36b7167fe80 req-eb1b0105-2ed7-4407-8e5c-ff408368721b service nova] [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] Refreshing instance network info cache due to event network-changed-a9945f05-55e7-4534-a38e-d921a091bb11. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 819.655938] env[63202]: DEBUG oslo_concurrency.lockutils [req-4709b7ef-80bb-4d16-80cb-a36b7167fe80 req-eb1b0105-2ed7-4407-8e5c-ff408368721b service nova] Acquiring lock "refresh_cache-39d9b91b-c009-40e1-ad2b-ebef650188a4" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.768864] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 8cd0415d-dab9-4554-966c-1ea8d822deff] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 819.830374] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.374s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.830906] env[63202]: DEBUG nova.compute.manager [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 819.833414] env[63202]: DEBUG oslo_concurrency.lockutils [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.575s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.834760] env[63202]: INFO nova.compute.claims [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 820.083637] env[63202]: DEBUG nova.network.neutron [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 820.158578] env[63202]: DEBUG nova.network.neutron [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.274398] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 21d0b283-3bb4-4bda-8e62-c933c6de4927] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 820.338586] env[63202]: DEBUG nova.compute.utils [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 820.341461] env[63202]: DEBUG nova.compute.manager [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 820.341624] env[63202]: DEBUG nova.network.neutron [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 820.398719] env[63202]: DEBUG nova.policy [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ba9bdc218e2b44b2a338e799a6fb84a0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fc7e68f0c5e3411a81971d90c03adf52', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 820.663756] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Releasing lock "refresh_cache-39d9b91b-c009-40e1-ad2b-ebef650188a4" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.663756] env[63202]: DEBUG nova.compute.manager [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 820.663756] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 820.663756] env[63202]: DEBUG oslo_concurrency.lockutils [req-4709b7ef-80bb-4d16-80cb-a36b7167fe80 req-eb1b0105-2ed7-4407-8e5c-ff408368721b service nova] Acquired lock "refresh_cache-39d9b91b-c009-40e1-ad2b-ebef650188a4" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.663756] env[63202]: DEBUG nova.network.neutron [req-4709b7ef-80bb-4d16-80cb-a36b7167fe80 req-eb1b0105-2ed7-4407-8e5c-ff408368721b service nova] [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] Refreshing network info cache for port a9945f05-55e7-4534-a38e-d921a091bb11 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 820.664188] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2337ae33-11d4-4076-ab0d-f6da351b4bc5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.676435] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23839fe1-5c8a-4ad9-9aa3-72c609476f66 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.696895] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 39d9b91b-c009-40e1-ad2b-ebef650188a4 could not be found. [ 820.697142] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 820.697340] env[63202]: INFO nova.compute.manager [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] Took 0.04 seconds to destroy the instance on the hypervisor. [ 820.697595] env[63202]: DEBUG oslo.service.loopingcall [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 820.697821] env[63202]: DEBUG nova.compute.manager [-] [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 820.697912] env[63202]: DEBUG nova.network.neutron [-] [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 820.713706] env[63202]: DEBUG nova.network.neutron [-] [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 820.723724] env[63202]: DEBUG nova.network.neutron [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] Successfully created port: eddbfe5e-fc77-4e94-9795-fb3364bdbd51 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 820.777309] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 0ded82aa-c68d-4d11-9e4d-90ed9a19b708] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 820.844869] env[63202]: DEBUG nova.compute.manager [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 821.094451] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f87879fe-3ce9-46c8-a0f5-7de862c55d1e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.102111] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47a59ef2-d61a-42cd-a53f-678ca9736069 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.131927] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd8a1dfd-44c6-414e-b529-034ad26c7cd6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.138540] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-723984f9-f5fc-469f-bdcf-863fbd2f9598 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.151137] env[63202]: DEBUG nova.compute.provider_tree [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 821.182710] env[63202]: DEBUG nova.network.neutron [req-4709b7ef-80bb-4d16-80cb-a36b7167fe80 req-eb1b0105-2ed7-4407-8e5c-ff408368721b service nova] [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 821.218179] env[63202]: DEBUG nova.network.neutron [-] [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.265033] env[63202]: DEBUG nova.network.neutron [req-4709b7ef-80bb-4d16-80cb-a36b7167fe80 req-eb1b0105-2ed7-4407-8e5c-ff408368721b service nova] [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.281950] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 821.282133] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Cleaning up deleted instances with incomplete migration {{(pid=63202) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 821.654325] env[63202]: DEBUG nova.scheduler.client.report [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 821.679762] env[63202]: DEBUG nova.compute.manager [req-679e8a50-aeab-4053-af03-88211d4d31df req-19ff917b-18c0-4b6e-bc63-7e5f59fd420d service nova] [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] Received event network-changed-eddbfe5e-fc77-4e94-9795-fb3364bdbd51 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 821.680021] env[63202]: DEBUG nova.compute.manager [req-679e8a50-aeab-4053-af03-88211d4d31df req-19ff917b-18c0-4b6e-bc63-7e5f59fd420d service nova] [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] Refreshing instance network info cache due to event network-changed-eddbfe5e-fc77-4e94-9795-fb3364bdbd51. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 821.680281] env[63202]: DEBUG oslo_concurrency.lockutils [req-679e8a50-aeab-4053-af03-88211d4d31df req-19ff917b-18c0-4b6e-bc63-7e5f59fd420d service nova] Acquiring lock "refresh_cache-06bd8147-5479-4c5e-8ba2-6d3a205ba05e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.680465] env[63202]: DEBUG oslo_concurrency.lockutils [req-679e8a50-aeab-4053-af03-88211d4d31df req-19ff917b-18c0-4b6e-bc63-7e5f59fd420d service nova] Acquired lock "refresh_cache-06bd8147-5479-4c5e-8ba2-6d3a205ba05e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.680623] env[63202]: DEBUG nova.network.neutron [req-679e8a50-aeab-4053-af03-88211d4d31df req-19ff917b-18c0-4b6e-bc63-7e5f59fd420d service nova] [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] Refreshing network info cache for port eddbfe5e-fc77-4e94-9795-fb3364bdbd51 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 821.719938] env[63202]: INFO nova.compute.manager [-] [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] Took 1.02 seconds to deallocate network for instance. [ 821.722655] env[63202]: DEBUG nova.compute.claims [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 821.722801] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.769105] env[63202]: DEBUG oslo_concurrency.lockutils [req-4709b7ef-80bb-4d16-80cb-a36b7167fe80 req-eb1b0105-2ed7-4407-8e5c-ff408368721b service nova] Releasing lock "refresh_cache-39d9b91b-c009-40e1-ad2b-ebef650188a4" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.769398] env[63202]: DEBUG nova.compute.manager [req-4709b7ef-80bb-4d16-80cb-a36b7167fe80 req-eb1b0105-2ed7-4407-8e5c-ff408368721b service nova] [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] Received event network-vif-deleted-a9945f05-55e7-4534-a38e-d921a091bb11 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 821.784748] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 821.804155] env[63202]: ERROR nova.compute.manager [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port eddbfe5e-fc77-4e94-9795-fb3364bdbd51, please check neutron logs for more information. [ 821.804155] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 821.804155] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 821.804155] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 821.804155] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 821.804155] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 821.804155] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 821.804155] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 821.804155] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 821.804155] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 821.804155] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 821.804155] env[63202]: ERROR nova.compute.manager raise self.value [ 821.804155] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 821.804155] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 821.804155] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 821.804155] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 821.804689] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 821.804689] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 821.804689] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port eddbfe5e-fc77-4e94-9795-fb3364bdbd51, please check neutron logs for more information. [ 821.804689] env[63202]: ERROR nova.compute.manager [ 821.804689] env[63202]: Traceback (most recent call last): [ 821.804689] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 821.804689] env[63202]: listener.cb(fileno) [ 821.804689] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 821.804689] env[63202]: result = function(*args, **kwargs) [ 821.804689] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 821.804689] env[63202]: return func(*args, **kwargs) [ 821.804689] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 821.804689] env[63202]: raise e [ 821.804689] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 821.804689] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 821.804689] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 821.804689] env[63202]: created_port_ids = self._update_ports_for_instance( [ 821.804689] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 821.804689] env[63202]: with excutils.save_and_reraise_exception(): [ 821.804689] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 821.804689] env[63202]: self.force_reraise() [ 821.804689] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 821.804689] env[63202]: raise self.value [ 821.804689] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 821.804689] env[63202]: updated_port = self._update_port( [ 821.804689] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 821.804689] env[63202]: _ensure_no_port_binding_failure(port) [ 821.804689] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 821.804689] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 821.805615] env[63202]: nova.exception.PortBindingFailed: Binding failed for port eddbfe5e-fc77-4e94-9795-fb3364bdbd51, please check neutron logs for more information. [ 821.805615] env[63202]: Removing descriptor: 16 [ 821.857150] env[63202]: DEBUG nova.compute.manager [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 821.881972] env[63202]: DEBUG nova.virt.hardware [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 821.882235] env[63202]: DEBUG nova.virt.hardware [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 821.882385] env[63202]: DEBUG nova.virt.hardware [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 821.882564] env[63202]: DEBUG nova.virt.hardware [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 821.882709] env[63202]: DEBUG nova.virt.hardware [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 821.882889] env[63202]: DEBUG nova.virt.hardware [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 821.883116] env[63202]: DEBUG nova.virt.hardware [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 821.883275] env[63202]: DEBUG nova.virt.hardware [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 821.883437] env[63202]: DEBUG nova.virt.hardware [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 821.883589] env[63202]: DEBUG nova.virt.hardware [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 821.883755] env[63202]: DEBUG nova.virt.hardware [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 821.884647] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b8056a0-6e67-4aa3-9196-1d77453b4c13 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.893071] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06ce25d6-981d-4518-beb1-91d66879d9da {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.907512] env[63202]: ERROR nova.compute.manager [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port eddbfe5e-fc77-4e94-9795-fb3364bdbd51, please check neutron logs for more information. [ 821.907512] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] Traceback (most recent call last): [ 821.907512] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 821.907512] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] yield resources [ 821.907512] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 821.907512] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] self.driver.spawn(context, instance, image_meta, [ 821.907512] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 821.907512] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 821.907512] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 821.907512] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] vm_ref = self.build_virtual_machine(instance, [ 821.907512] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 821.907975] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] vif_infos = vmwarevif.get_vif_info(self._session, [ 821.907975] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 821.907975] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] for vif in network_info: [ 821.907975] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 821.907975] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] return self._sync_wrapper(fn, *args, **kwargs) [ 821.907975] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 821.907975] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] self.wait() [ 821.907975] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 821.907975] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] self[:] = self._gt.wait() [ 821.907975] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 821.907975] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] return self._exit_event.wait() [ 821.907975] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 821.907975] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] current.throw(*self._exc) [ 821.908563] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 821.908563] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] result = function(*args, **kwargs) [ 821.908563] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 821.908563] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] return func(*args, **kwargs) [ 821.908563] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 821.908563] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] raise e [ 821.908563] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 821.908563] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] nwinfo = self.network_api.allocate_for_instance( [ 821.908563] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 821.908563] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] created_port_ids = self._update_ports_for_instance( [ 821.908563] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 821.908563] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] with excutils.save_and_reraise_exception(): [ 821.908563] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 821.909052] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] self.force_reraise() [ 821.909052] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 821.909052] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] raise self.value [ 821.909052] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 821.909052] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] updated_port = self._update_port( [ 821.909052] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 821.909052] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] _ensure_no_port_binding_failure(port) [ 821.909052] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 821.909052] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] raise exception.PortBindingFailed(port_id=port['id']) [ 821.909052] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] nova.exception.PortBindingFailed: Binding failed for port eddbfe5e-fc77-4e94-9795-fb3364bdbd51, please check neutron logs for more information. [ 821.909052] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] [ 821.909052] env[63202]: INFO nova.compute.manager [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] Terminating instance [ 821.911257] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Acquiring lock "refresh_cache-06bd8147-5479-4c5e-8ba2-6d3a205ba05e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.161317] env[63202]: DEBUG oslo_concurrency.lockutils [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.328s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.161823] env[63202]: DEBUG nova.compute.manager [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 822.164419] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.035s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.165846] env[63202]: INFO nova.compute.claims [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 822.197211] env[63202]: DEBUG nova.network.neutron [req-679e8a50-aeab-4053-af03-88211d4d31df req-19ff917b-18c0-4b6e-bc63-7e5f59fd420d service nova] [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 822.282136] env[63202]: DEBUG nova.network.neutron [req-679e8a50-aeab-4053-af03-88211d4d31df req-19ff917b-18c0-4b6e-bc63-7e5f59fd420d service nova] [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.670616] env[63202]: DEBUG nova.compute.utils [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 822.673868] env[63202]: DEBUG nova.compute.manager [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 822.674041] env[63202]: DEBUG nova.network.neutron [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 822.737460] env[63202]: DEBUG nova.policy [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '67da6820f3cd4db987724a24b7ae1fc6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'be0c9c1bdfaf433bb99d7a9f6de0a643', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 822.785041] env[63202]: DEBUG oslo_concurrency.lockutils [req-679e8a50-aeab-4053-af03-88211d4d31df req-19ff917b-18c0-4b6e-bc63-7e5f59fd420d service nova] Releasing lock "refresh_cache-06bd8147-5479-4c5e-8ba2-6d3a205ba05e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.785488] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Acquired lock "refresh_cache-06bd8147-5479-4c5e-8ba2-6d3a205ba05e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.785673] env[63202]: DEBUG nova.network.neutron [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 823.147134] env[63202]: DEBUG nova.network.neutron [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] Successfully created port: a89d882c-6b78-4f47-929d-8d342776d561 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 823.175200] env[63202]: DEBUG nova.compute.manager [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 823.288028] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 823.290118] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 823.290339] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 823.290768] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 823.290938] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 823.291196] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63202) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 823.291325] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager.update_available_resource {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 823.307343] env[63202]: DEBUG nova.network.neutron [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 823.384216] env[63202]: DEBUG nova.network.neutron [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.458184] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-109040c7-7963-4d4a-bdd6-527f5a90778a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.465902] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbc30a82-7267-4366-ba29-36d0fd486606 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.495389] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0ac3f98-cdb2-43b2-b47d-40dfdf8a3e4a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.502349] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-155ae446-e6a5-402e-b262-7ecb5c2d0e2e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.514933] env[63202]: DEBUG nova.compute.provider_tree [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 823.745597] env[63202]: DEBUG nova.compute.manager [req-aa92367b-f5f6-4cb4-9faf-43e4f99d4029 req-748a8ff0-eb14-42e8-991e-483676c99533 service nova] [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] Received event network-vif-deleted-eddbfe5e-fc77-4e94-9795-fb3364bdbd51 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 823.797362] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.894073] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Releasing lock "refresh_cache-06bd8147-5479-4c5e-8ba2-6d3a205ba05e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.894177] env[63202]: DEBUG nova.compute.manager [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 823.894369] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 823.894946] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-53018a7e-acac-48cc-b631-a670206ac2a5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.904332] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82929f7d-16cf-4f01-a27e-513c0f1c9fc4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.928022] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 06bd8147-5479-4c5e-8ba2-6d3a205ba05e could not be found. [ 823.928262] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 823.928437] env[63202]: INFO nova.compute.manager [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] Took 0.03 seconds to destroy the instance on the hypervisor. [ 823.928673] env[63202]: DEBUG oslo.service.loopingcall [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 823.930824] env[63202]: DEBUG nova.compute.manager [-] [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 823.930824] env[63202]: DEBUG nova.network.neutron [-] [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 823.945937] env[63202]: DEBUG nova.network.neutron [-] [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 824.018547] env[63202]: DEBUG nova.scheduler.client.report [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 824.187749] env[63202]: DEBUG nova.compute.manager [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 824.214074] env[63202]: DEBUG nova.virt.hardware [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 824.214322] env[63202]: DEBUG nova.virt.hardware [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 824.214477] env[63202]: DEBUG nova.virt.hardware [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 824.214653] env[63202]: DEBUG nova.virt.hardware [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 824.214850] env[63202]: DEBUG nova.virt.hardware [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 824.215018] env[63202]: DEBUG nova.virt.hardware [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 824.215742] env[63202]: DEBUG nova.virt.hardware [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 824.215963] env[63202]: DEBUG nova.virt.hardware [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 824.216166] env[63202]: DEBUG nova.virt.hardware [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 824.216335] env[63202]: DEBUG nova.virt.hardware [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 824.217033] env[63202]: DEBUG nova.virt.hardware [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 824.218023] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33657c17-d635-41d1-9aea-26e9101762df {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.225902] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62cfee72-39f6-4290-b2ec-d135ebe2e11e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.300736] env[63202]: ERROR nova.compute.manager [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a89d882c-6b78-4f47-929d-8d342776d561, please check neutron logs for more information. [ 824.300736] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 824.300736] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 824.300736] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 824.300736] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 824.300736] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 824.300736] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 824.300736] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 824.300736] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 824.300736] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 824.300736] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 824.300736] env[63202]: ERROR nova.compute.manager raise self.value [ 824.300736] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 824.300736] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 824.300736] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 824.300736] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 824.301290] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 824.301290] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 824.301290] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a89d882c-6b78-4f47-929d-8d342776d561, please check neutron logs for more information. [ 824.301290] env[63202]: ERROR nova.compute.manager [ 824.301290] env[63202]: Traceback (most recent call last): [ 824.301290] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 824.301290] env[63202]: listener.cb(fileno) [ 824.301290] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 824.301290] env[63202]: result = function(*args, **kwargs) [ 824.301290] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 824.301290] env[63202]: return func(*args, **kwargs) [ 824.301290] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 824.301290] env[63202]: raise e [ 824.301290] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 824.301290] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 824.301290] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 824.301290] env[63202]: created_port_ids = self._update_ports_for_instance( [ 824.301290] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 824.301290] env[63202]: with excutils.save_and_reraise_exception(): [ 824.301290] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 824.301290] env[63202]: self.force_reraise() [ 824.301290] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 824.301290] env[63202]: raise self.value [ 824.301290] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 824.301290] env[63202]: updated_port = self._update_port( [ 824.301290] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 824.301290] env[63202]: _ensure_no_port_binding_failure(port) [ 824.301290] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 824.301290] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 824.302222] env[63202]: nova.exception.PortBindingFailed: Binding failed for port a89d882c-6b78-4f47-929d-8d342776d561, please check neutron logs for more information. [ 824.302222] env[63202]: Removing descriptor: 16 [ 824.302222] env[63202]: ERROR nova.compute.manager [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a89d882c-6b78-4f47-929d-8d342776d561, please check neutron logs for more information. [ 824.302222] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] Traceback (most recent call last): [ 824.302222] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 824.302222] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] yield resources [ 824.302222] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 824.302222] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] self.driver.spawn(context, instance, image_meta, [ 824.302222] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 824.302222] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 824.302222] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 824.302222] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] vm_ref = self.build_virtual_machine(instance, [ 824.302622] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 824.302622] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] vif_infos = vmwarevif.get_vif_info(self._session, [ 824.302622] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 824.302622] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] for vif in network_info: [ 824.302622] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 824.302622] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] return self._sync_wrapper(fn, *args, **kwargs) [ 824.302622] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 824.302622] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] self.wait() [ 824.302622] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 824.302622] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] self[:] = self._gt.wait() [ 824.302622] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 824.302622] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] return self._exit_event.wait() [ 824.302622] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 824.303070] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] result = hub.switch() [ 824.303070] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 824.303070] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] return self.greenlet.switch() [ 824.303070] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 824.303070] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] result = function(*args, **kwargs) [ 824.303070] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 824.303070] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] return func(*args, **kwargs) [ 824.303070] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 824.303070] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] raise e [ 824.303070] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 824.303070] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] nwinfo = self.network_api.allocate_for_instance( [ 824.303070] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 824.303070] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] created_port_ids = self._update_ports_for_instance( [ 824.303505] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 824.303505] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] with excutils.save_and_reraise_exception(): [ 824.303505] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 824.303505] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] self.force_reraise() [ 824.303505] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 824.303505] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] raise self.value [ 824.303505] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 824.303505] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] updated_port = self._update_port( [ 824.303505] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 824.303505] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] _ensure_no_port_binding_failure(port) [ 824.303505] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 824.303505] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] raise exception.PortBindingFailed(port_id=port['id']) [ 824.303874] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] nova.exception.PortBindingFailed: Binding failed for port a89d882c-6b78-4f47-929d-8d342776d561, please check neutron logs for more information. [ 824.303874] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] [ 824.303874] env[63202]: INFO nova.compute.manager [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] Terminating instance [ 824.303980] env[63202]: DEBUG oslo_concurrency.lockutils [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] Acquiring lock "refresh_cache-56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.304057] env[63202]: DEBUG oslo_concurrency.lockutils [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] Acquired lock "refresh_cache-56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.304220] env[63202]: DEBUG nova.network.neutron [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 824.448504] env[63202]: DEBUG nova.network.neutron [-] [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.523957] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.359s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.524525] env[63202]: DEBUG nova.compute.manager [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 824.527704] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.483s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.529209] env[63202]: INFO nova.compute.claims [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 824.819736] env[63202]: DEBUG nova.network.neutron [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 824.866136] env[63202]: DEBUG nova.network.neutron [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.952332] env[63202]: INFO nova.compute.manager [-] [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] Took 1.02 seconds to deallocate network for instance. [ 824.956296] env[63202]: DEBUG nova.compute.claims [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 824.956476] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.029580] env[63202]: DEBUG nova.compute.utils [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 825.030973] env[63202]: DEBUG nova.compute.manager [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 825.031136] env[63202]: DEBUG nova.network.neutron [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 825.091118] env[63202]: DEBUG nova.policy [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1236175893a3401cb01b04604168f1e0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '22352c25bca8416a948014391a5389ac', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 825.370434] env[63202]: DEBUG oslo_concurrency.lockutils [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] Releasing lock "refresh_cache-56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.370874] env[63202]: DEBUG nova.compute.manager [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 825.371083] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 825.371387] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5cdc8646-70bd-4dbb-82a4-24b3ef861419 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.380647] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cab7ca93-eb1f-4731-8059-d9950ee09407 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.402257] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a could not be found. [ 825.402518] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 825.402703] env[63202]: INFO nova.compute.manager [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] Took 0.03 seconds to destroy the instance on the hypervisor. [ 825.402934] env[63202]: DEBUG oslo.service.loopingcall [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 825.403157] env[63202]: DEBUG nova.compute.manager [-] [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 825.403245] env[63202]: DEBUG nova.network.neutron [-] [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 825.415257] env[63202]: DEBUG nova.network.neutron [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] Successfully created port: 115c9b60-ce54-4ffa-9bb5-eae100911988 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 825.417710] env[63202]: DEBUG nova.network.neutron [-] [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 825.535297] env[63202]: DEBUG nova.compute.manager [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 825.770312] env[63202]: DEBUG nova.compute.manager [req-85cf91b0-8e5a-4fe9-9da0-6364a2e436f7 req-69ffecab-8cca-4368-8199-0b9d5ea587f5 service nova] [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] Received event network-changed-a89d882c-6b78-4f47-929d-8d342776d561 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 825.770573] env[63202]: DEBUG nova.compute.manager [req-85cf91b0-8e5a-4fe9-9da0-6364a2e436f7 req-69ffecab-8cca-4368-8199-0b9d5ea587f5 service nova] [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] Refreshing instance network info cache due to event network-changed-a89d882c-6b78-4f47-929d-8d342776d561. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 825.770695] env[63202]: DEBUG oslo_concurrency.lockutils [req-85cf91b0-8e5a-4fe9-9da0-6364a2e436f7 req-69ffecab-8cca-4368-8199-0b9d5ea587f5 service nova] Acquiring lock "refresh_cache-56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.770960] env[63202]: DEBUG oslo_concurrency.lockutils [req-85cf91b0-8e5a-4fe9-9da0-6364a2e436f7 req-69ffecab-8cca-4368-8199-0b9d5ea587f5 service nova] Acquired lock "refresh_cache-56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.770960] env[63202]: DEBUG nova.network.neutron [req-85cf91b0-8e5a-4fe9-9da0-6364a2e436f7 req-69ffecab-8cca-4368-8199-0b9d5ea587f5 service nova] [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] Refreshing network info cache for port a89d882c-6b78-4f47-929d-8d342776d561 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 825.774231] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39d710a2-ae4e-42c9-9bc0-c2d1e9405bce {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.782107] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c780d508-6f01-455b-8d7d-f55882a87d86 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.812643] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4f7ba9e-def4-4d50-aeb2-ee7a1854f4f2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.819849] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-700f54d0-ca22-445e-b8bc-6e76951d8e10 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.833974] env[63202]: DEBUG nova.compute.provider_tree [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 825.920589] env[63202]: DEBUG nova.network.neutron [-] [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.293730] env[63202]: DEBUG nova.network.neutron [req-85cf91b0-8e5a-4fe9-9da0-6364a2e436f7 req-69ffecab-8cca-4368-8199-0b9d5ea587f5 service nova] [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 826.338327] env[63202]: DEBUG nova.scheduler.client.report [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 826.346027] env[63202]: DEBUG nova.network.neutron [req-85cf91b0-8e5a-4fe9-9da0-6364a2e436f7 req-69ffecab-8cca-4368-8199-0b9d5ea587f5 service nova] [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.423962] env[63202]: INFO nova.compute.manager [-] [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] Took 1.02 seconds to deallocate network for instance. [ 826.426256] env[63202]: DEBUG nova.compute.claims [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 826.426428] env[63202]: DEBUG oslo_concurrency.lockutils [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.546255] env[63202]: DEBUG nova.compute.manager [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 826.575047] env[63202]: DEBUG nova.virt.hardware [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 826.575301] env[63202]: DEBUG nova.virt.hardware [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 826.575476] env[63202]: DEBUG nova.virt.hardware [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 826.575637] env[63202]: DEBUG nova.virt.hardware [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 826.575779] env[63202]: DEBUG nova.virt.hardware [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 826.575919] env[63202]: DEBUG nova.virt.hardware [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 826.576143] env[63202]: DEBUG nova.virt.hardware [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 826.576299] env[63202]: DEBUG nova.virt.hardware [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 826.576450] env[63202]: DEBUG nova.virt.hardware [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 826.576615] env[63202]: DEBUG nova.virt.hardware [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 826.576787] env[63202]: DEBUG nova.virt.hardware [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 826.577669] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae126d4d-1c6a-4f88-82d8-54c3a0d60126 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.585677] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee3a2d84-e8e5-42ed-b292-616c4c2e492b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.650683] env[63202]: ERROR nova.compute.manager [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 115c9b60-ce54-4ffa-9bb5-eae100911988, please check neutron logs for more information. [ 826.650683] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 826.650683] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 826.650683] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 826.650683] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 826.650683] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 826.650683] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 826.650683] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 826.650683] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 826.650683] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 826.650683] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 826.650683] env[63202]: ERROR nova.compute.manager raise self.value [ 826.650683] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 826.650683] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 826.650683] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 826.650683] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 826.651248] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 826.651248] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 826.651248] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 115c9b60-ce54-4ffa-9bb5-eae100911988, please check neutron logs for more information. [ 826.651248] env[63202]: ERROR nova.compute.manager [ 826.651248] env[63202]: Traceback (most recent call last): [ 826.651248] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 826.651248] env[63202]: listener.cb(fileno) [ 826.651248] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 826.651248] env[63202]: result = function(*args, **kwargs) [ 826.651248] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 826.651248] env[63202]: return func(*args, **kwargs) [ 826.651248] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 826.651248] env[63202]: raise e [ 826.651248] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 826.651248] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 826.651248] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 826.651248] env[63202]: created_port_ids = self._update_ports_for_instance( [ 826.651248] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 826.651248] env[63202]: with excutils.save_and_reraise_exception(): [ 826.651248] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 826.651248] env[63202]: self.force_reraise() [ 826.651248] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 826.651248] env[63202]: raise self.value [ 826.651248] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 826.651248] env[63202]: updated_port = self._update_port( [ 826.651248] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 826.651248] env[63202]: _ensure_no_port_binding_failure(port) [ 826.651248] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 826.651248] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 826.652175] env[63202]: nova.exception.PortBindingFailed: Binding failed for port 115c9b60-ce54-4ffa-9bb5-eae100911988, please check neutron logs for more information. [ 826.652175] env[63202]: Removing descriptor: 16 [ 826.652175] env[63202]: ERROR nova.compute.manager [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 115c9b60-ce54-4ffa-9bb5-eae100911988, please check neutron logs for more information. [ 826.652175] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] Traceback (most recent call last): [ 826.652175] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 826.652175] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] yield resources [ 826.652175] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 826.652175] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] self.driver.spawn(context, instance, image_meta, [ 826.652175] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 826.652175] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 826.652175] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 826.652175] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] vm_ref = self.build_virtual_machine(instance, [ 826.652576] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 826.652576] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] vif_infos = vmwarevif.get_vif_info(self._session, [ 826.652576] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 826.652576] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] for vif in network_info: [ 826.652576] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 826.652576] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] return self._sync_wrapper(fn, *args, **kwargs) [ 826.652576] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 826.652576] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] self.wait() [ 826.652576] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 826.652576] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] self[:] = self._gt.wait() [ 826.652576] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 826.652576] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] return self._exit_event.wait() [ 826.652576] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 826.652994] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] result = hub.switch() [ 826.652994] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 826.652994] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] return self.greenlet.switch() [ 826.652994] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 826.652994] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] result = function(*args, **kwargs) [ 826.652994] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 826.652994] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] return func(*args, **kwargs) [ 826.652994] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 826.652994] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] raise e [ 826.652994] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 826.652994] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] nwinfo = self.network_api.allocate_for_instance( [ 826.652994] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 826.652994] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] created_port_ids = self._update_ports_for_instance( [ 826.653426] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 826.653426] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] with excutils.save_and_reraise_exception(): [ 826.653426] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 826.653426] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] self.force_reraise() [ 826.653426] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 826.653426] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] raise self.value [ 826.653426] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 826.653426] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] updated_port = self._update_port( [ 826.653426] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 826.653426] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] _ensure_no_port_binding_failure(port) [ 826.653426] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 826.653426] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] raise exception.PortBindingFailed(port_id=port['id']) [ 826.653814] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] nova.exception.PortBindingFailed: Binding failed for port 115c9b60-ce54-4ffa-9bb5-eae100911988, please check neutron logs for more information. [ 826.653814] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] [ 826.653814] env[63202]: INFO nova.compute.manager [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] Terminating instance [ 826.654675] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "refresh_cache-0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.654854] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquired lock "refresh_cache-0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.655033] env[63202]: DEBUG nova.network.neutron [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 826.843472] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.316s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.843991] env[63202]: DEBUG nova.compute.manager [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 826.847052] env[63202]: DEBUG oslo_concurrency.lockutils [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.977s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.848507] env[63202]: INFO nova.compute.claims [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 826.851149] env[63202]: DEBUG oslo_concurrency.lockutils [req-85cf91b0-8e5a-4fe9-9da0-6364a2e436f7 req-69ffecab-8cca-4368-8199-0b9d5ea587f5 service nova] Releasing lock "refresh_cache-56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.851359] env[63202]: DEBUG nova.compute.manager [req-85cf91b0-8e5a-4fe9-9da0-6364a2e436f7 req-69ffecab-8cca-4368-8199-0b9d5ea587f5 service nova] [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] Received event network-vif-deleted-a89d882c-6b78-4f47-929d-8d342776d561 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 827.177969] env[63202]: DEBUG nova.network.neutron [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 827.258791] env[63202]: DEBUG nova.network.neutron [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.353042] env[63202]: DEBUG nova.compute.utils [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 827.357106] env[63202]: DEBUG nova.compute.manager [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 827.357331] env[63202]: DEBUG nova.network.neutron [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 827.407550] env[63202]: DEBUG nova.policy [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a5b0e32aa8974a2c8556f8c50c807b99', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9a0d7d33a81d42428d5e8aeab0b90fa0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 827.755609] env[63202]: DEBUG nova.network.neutron [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] Successfully created port: 9f87d11f-db0c-4b6b-84a2-86c00886f0db {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 827.762190] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Releasing lock "refresh_cache-0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.762592] env[63202]: DEBUG nova.compute.manager [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 827.762775] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 827.763065] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0eaf0603-5d31-4fa1-9dc1-d31fd6d6a0e9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.772427] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d728cb7-e15e-4857-a364-71e17a66318f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.793988] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7 could not be found. [ 827.794219] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 827.794399] env[63202]: INFO nova.compute.manager [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] Took 0.03 seconds to destroy the instance on the hypervisor. [ 827.794653] env[63202]: DEBUG oslo.service.loopingcall [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 827.796061] env[63202]: DEBUG nova.compute.manager [-] [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 827.796061] env[63202]: DEBUG nova.network.neutron [-] [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 827.798386] env[63202]: DEBUG nova.compute.manager [req-56c71b1a-1522-420c-b49d-1c50ca11556c req-39d0da75-733a-4f47-96e0-536f45b0191a service nova] [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] Received event network-changed-115c9b60-ce54-4ffa-9bb5-eae100911988 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 827.798611] env[63202]: DEBUG nova.compute.manager [req-56c71b1a-1522-420c-b49d-1c50ca11556c req-39d0da75-733a-4f47-96e0-536f45b0191a service nova] [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] Refreshing instance network info cache due to event network-changed-115c9b60-ce54-4ffa-9bb5-eae100911988. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 827.798755] env[63202]: DEBUG oslo_concurrency.lockutils [req-56c71b1a-1522-420c-b49d-1c50ca11556c req-39d0da75-733a-4f47-96e0-536f45b0191a service nova] Acquiring lock "refresh_cache-0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.798946] env[63202]: DEBUG oslo_concurrency.lockutils [req-56c71b1a-1522-420c-b49d-1c50ca11556c req-39d0da75-733a-4f47-96e0-536f45b0191a service nova] Acquired lock "refresh_cache-0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.799147] env[63202]: DEBUG nova.network.neutron [req-56c71b1a-1522-420c-b49d-1c50ca11556c req-39d0da75-733a-4f47-96e0-536f45b0191a service nova] [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] Refreshing network info cache for port 115c9b60-ce54-4ffa-9bb5-eae100911988 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 827.815271] env[63202]: DEBUG nova.network.neutron [-] [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 827.860560] env[63202]: DEBUG nova.compute.manager [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 828.276110] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2af0fc67-6d90-4003-81df-bb93928bcdda {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.283725] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd1957c8-6456-4c09-83de-993fdd16c4f3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.315433] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64efdb25-53bf-4093-952c-56671bbbde2b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.318180] env[63202]: DEBUG nova.network.neutron [-] [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.328279] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fca230b-7795-47a0-bd3e-f04d22c83631 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.339579] env[63202]: DEBUG nova.compute.provider_tree [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 828.344020] env[63202]: DEBUG nova.network.neutron [req-56c71b1a-1522-420c-b49d-1c50ca11556c req-39d0da75-733a-4f47-96e0-536f45b0191a service nova] [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 828.421525] env[63202]: DEBUG nova.network.neutron [req-56c71b1a-1522-420c-b49d-1c50ca11556c req-39d0da75-733a-4f47-96e0-536f45b0191a service nova] [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.794929] env[63202]: ERROR nova.compute.manager [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9f87d11f-db0c-4b6b-84a2-86c00886f0db, please check neutron logs for more information. [ 828.794929] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 828.794929] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 828.794929] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 828.794929] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 828.794929] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 828.794929] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 828.794929] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 828.794929] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 828.794929] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 828.794929] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 828.794929] env[63202]: ERROR nova.compute.manager raise self.value [ 828.794929] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 828.794929] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 828.794929] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 828.794929] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 828.795472] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 828.795472] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 828.795472] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9f87d11f-db0c-4b6b-84a2-86c00886f0db, please check neutron logs for more information. [ 828.795472] env[63202]: ERROR nova.compute.manager [ 828.795472] env[63202]: Traceback (most recent call last): [ 828.795472] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 828.795472] env[63202]: listener.cb(fileno) [ 828.795472] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 828.795472] env[63202]: result = function(*args, **kwargs) [ 828.795472] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 828.795472] env[63202]: return func(*args, **kwargs) [ 828.795472] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 828.795472] env[63202]: raise e [ 828.795472] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 828.795472] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 828.795472] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 828.795472] env[63202]: created_port_ids = self._update_ports_for_instance( [ 828.795472] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 828.795472] env[63202]: with excutils.save_and_reraise_exception(): [ 828.795472] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 828.795472] env[63202]: self.force_reraise() [ 828.795472] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 828.795472] env[63202]: raise self.value [ 828.795472] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 828.795472] env[63202]: updated_port = self._update_port( [ 828.795472] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 828.795472] env[63202]: _ensure_no_port_binding_failure(port) [ 828.795472] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 828.795472] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 828.796401] env[63202]: nova.exception.PortBindingFailed: Binding failed for port 9f87d11f-db0c-4b6b-84a2-86c00886f0db, please check neutron logs for more information. [ 828.796401] env[63202]: Removing descriptor: 16 [ 828.824822] env[63202]: INFO nova.compute.manager [-] [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] Took 1.03 seconds to deallocate network for instance. [ 828.827078] env[63202]: DEBUG nova.compute.claims [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 828.827396] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.844648] env[63202]: DEBUG nova.scheduler.client.report [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 828.872331] env[63202]: DEBUG nova.compute.manager [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 828.898538] env[63202]: DEBUG nova.virt.hardware [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 828.898797] env[63202]: DEBUG nova.virt.hardware [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 828.898948] env[63202]: DEBUG nova.virt.hardware [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 828.899139] env[63202]: DEBUG nova.virt.hardware [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 828.899280] env[63202]: DEBUG nova.virt.hardware [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 828.899449] env[63202]: DEBUG nova.virt.hardware [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 828.899659] env[63202]: DEBUG nova.virt.hardware [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 828.899814] env[63202]: DEBUG nova.virt.hardware [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 828.899976] env[63202]: DEBUG nova.virt.hardware [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 828.900146] env[63202]: DEBUG nova.virt.hardware [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 828.900315] env[63202]: DEBUG nova.virt.hardware [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 828.901173] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a03c8f8e-9876-433d-a369-419f3efc70ad {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.908809] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42760728-289d-4fcf-bd3c-c02aa38f7478 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.922365] env[63202]: ERROR nova.compute.manager [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9f87d11f-db0c-4b6b-84a2-86c00886f0db, please check neutron logs for more information. [ 828.922365] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] Traceback (most recent call last): [ 828.922365] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 828.922365] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] yield resources [ 828.922365] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 828.922365] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] self.driver.spawn(context, instance, image_meta, [ 828.922365] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 828.922365] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 828.922365] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 828.922365] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] vm_ref = self.build_virtual_machine(instance, [ 828.922365] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 828.922728] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] vif_infos = vmwarevif.get_vif_info(self._session, [ 828.922728] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 828.922728] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] for vif in network_info: [ 828.922728] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 828.922728] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] return self._sync_wrapper(fn, *args, **kwargs) [ 828.922728] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 828.922728] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] self.wait() [ 828.922728] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 828.922728] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] self[:] = self._gt.wait() [ 828.922728] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 828.922728] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] return self._exit_event.wait() [ 828.922728] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 828.922728] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] current.throw(*self._exc) [ 828.923165] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 828.923165] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] result = function(*args, **kwargs) [ 828.923165] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 828.923165] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] return func(*args, **kwargs) [ 828.923165] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 828.923165] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] raise e [ 828.923165] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 828.923165] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] nwinfo = self.network_api.allocate_for_instance( [ 828.923165] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 828.923165] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] created_port_ids = self._update_ports_for_instance( [ 828.923165] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 828.923165] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] with excutils.save_and_reraise_exception(): [ 828.923165] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 828.923550] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] self.force_reraise() [ 828.923550] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 828.923550] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] raise self.value [ 828.923550] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 828.923550] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] updated_port = self._update_port( [ 828.923550] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 828.923550] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] _ensure_no_port_binding_failure(port) [ 828.923550] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 828.923550] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] raise exception.PortBindingFailed(port_id=port['id']) [ 828.923550] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] nova.exception.PortBindingFailed: Binding failed for port 9f87d11f-db0c-4b6b-84a2-86c00886f0db, please check neutron logs for more information. [ 828.923550] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] [ 828.923550] env[63202]: INFO nova.compute.manager [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] Terminating instance [ 828.924277] env[63202]: DEBUG oslo_concurrency.lockutils [req-56c71b1a-1522-420c-b49d-1c50ca11556c req-39d0da75-733a-4f47-96e0-536f45b0191a service nova] Releasing lock "refresh_cache-0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.924500] env[63202]: DEBUG nova.compute.manager [req-56c71b1a-1522-420c-b49d-1c50ca11556c req-39d0da75-733a-4f47-96e0-536f45b0191a service nova] [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] Received event network-vif-deleted-115c9b60-ce54-4ffa-9bb5-eae100911988 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 828.925103] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] Acquiring lock "refresh_cache-a47dbb3e-62be-426d-acd7-bedc0faf65c5" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.925258] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] Acquired lock "refresh_cache-a47dbb3e-62be-426d-acd7-bedc0faf65c5" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.925414] env[63202]: DEBUG nova.network.neutron [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 829.349165] env[63202]: DEBUG oslo_concurrency.lockutils [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.502s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.349747] env[63202]: DEBUG nova.compute.manager [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 829.352856] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.247s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.354228] env[63202]: INFO nova.compute.claims [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 829.443625] env[63202]: DEBUG nova.network.neutron [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 829.536338] env[63202]: DEBUG nova.network.neutron [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.819720] env[63202]: DEBUG nova.compute.manager [req-bd6111a1-7407-4951-bc48-d3e3dbd1e85b req-6a866ac2-00d7-45f3-8ee8-6644cc30898a service nova] [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] Received event network-changed-9f87d11f-db0c-4b6b-84a2-86c00886f0db {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 829.819918] env[63202]: DEBUG nova.compute.manager [req-bd6111a1-7407-4951-bc48-d3e3dbd1e85b req-6a866ac2-00d7-45f3-8ee8-6644cc30898a service nova] [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] Refreshing instance network info cache due to event network-changed-9f87d11f-db0c-4b6b-84a2-86c00886f0db. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 829.820114] env[63202]: DEBUG oslo_concurrency.lockutils [req-bd6111a1-7407-4951-bc48-d3e3dbd1e85b req-6a866ac2-00d7-45f3-8ee8-6644cc30898a service nova] Acquiring lock "refresh_cache-a47dbb3e-62be-426d-acd7-bedc0faf65c5" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.858105] env[63202]: DEBUG nova.compute.utils [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 829.861942] env[63202]: DEBUG nova.compute.manager [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 829.862172] env[63202]: DEBUG nova.network.neutron [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 829.900754] env[63202]: DEBUG nova.policy [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '46863ba14f1c4829a24eca823e763f2f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b028e34225744668807aae95712a8d41', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 830.039377] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] Releasing lock "refresh_cache-a47dbb3e-62be-426d-acd7-bedc0faf65c5" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.039758] env[63202]: DEBUG nova.compute.manager [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 830.039946] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 830.042088] env[63202]: DEBUG oslo_concurrency.lockutils [req-bd6111a1-7407-4951-bc48-d3e3dbd1e85b req-6a866ac2-00d7-45f3-8ee8-6644cc30898a service nova] Acquired lock "refresh_cache-a47dbb3e-62be-426d-acd7-bedc0faf65c5" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.042278] env[63202]: DEBUG nova.network.neutron [req-bd6111a1-7407-4951-bc48-d3e3dbd1e85b req-6a866ac2-00d7-45f3-8ee8-6644cc30898a service nova] [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] Refreshing network info cache for port 9f87d11f-db0c-4b6b-84a2-86c00886f0db {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 830.043300] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ad659f62-681c-407c-8fd9-aedf422a69a2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.052867] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1549aa28-6dd5-497a-b92b-3101f29ce3af {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.074229] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a47dbb3e-62be-426d-acd7-bedc0faf65c5 could not be found. [ 830.074420] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 830.074594] env[63202]: INFO nova.compute.manager [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] Took 0.03 seconds to destroy the instance on the hypervisor. [ 830.074824] env[63202]: DEBUG oslo.service.loopingcall [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 830.075039] env[63202]: DEBUG nova.compute.manager [-] [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 830.075133] env[63202]: DEBUG nova.network.neutron [-] [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 830.090589] env[63202]: DEBUG nova.network.neutron [-] [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 830.205227] env[63202]: DEBUG nova.network.neutron [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] Successfully created port: cc6896d9-1f04-49c5-82f3-1d50d3ef5244 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 830.368239] env[63202]: DEBUG nova.compute.manager [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 830.388320] env[63202]: DEBUG nova.scheduler.client.report [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] Refreshing inventories for resource provider 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 830.406333] env[63202]: DEBUG nova.scheduler.client.report [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] Updating ProviderTree inventory for provider 79b33d17-3e75-494c-a550-67b275de2079 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 830.406333] env[63202]: DEBUG nova.compute.provider_tree [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] Updating inventory in ProviderTree for provider 79b33d17-3e75-494c-a550-67b275de2079 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 830.417472] env[63202]: DEBUG nova.scheduler.client.report [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] Refreshing aggregate associations for resource provider 79b33d17-3e75-494c-a550-67b275de2079, aggregates: None {{(pid=63202) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 830.436188] env[63202]: DEBUG nova.scheduler.client.report [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] Refreshing trait associations for resource provider 79b33d17-3e75-494c-a550-67b275de2079, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,HW_ARCH_X86_64,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO {{(pid=63202) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 830.566929] env[63202]: DEBUG nova.network.neutron [req-bd6111a1-7407-4951-bc48-d3e3dbd1e85b req-6a866ac2-00d7-45f3-8ee8-6644cc30898a service nova] [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 830.593929] env[63202]: DEBUG nova.network.neutron [-] [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.668446] env[63202]: DEBUG nova.network.neutron [req-bd6111a1-7407-4951-bc48-d3e3dbd1e85b req-6a866ac2-00d7-45f3-8ee8-6644cc30898a service nova] [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.748896] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dd7b176-de16-46db-b5e7-5765a1fa8e00 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.757040] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c31e8707-6d77-434f-8bf4-2affe391634f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.793259] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e2022e1-b436-42d6-8e3e-4b2eec5addae {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.801177] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed7b1abc-a73c-4a84-9a56-0dd04658ea86 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.814822] env[63202]: DEBUG nova.compute.provider_tree [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 831.098030] env[63202]: INFO nova.compute.manager [-] [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] Took 1.02 seconds to deallocate network for instance. [ 831.100082] env[63202]: DEBUG nova.compute.claims [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 831.100082] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.173025] env[63202]: DEBUG oslo_concurrency.lockutils [req-bd6111a1-7407-4951-bc48-d3e3dbd1e85b req-6a866ac2-00d7-45f3-8ee8-6644cc30898a service nova] Releasing lock "refresh_cache-a47dbb3e-62be-426d-acd7-bedc0faf65c5" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.173275] env[63202]: DEBUG nova.compute.manager [req-bd6111a1-7407-4951-bc48-d3e3dbd1e85b req-6a866ac2-00d7-45f3-8ee8-6644cc30898a service nova] [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] Received event network-vif-deleted-9f87d11f-db0c-4b6b-84a2-86c00886f0db {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 831.318703] env[63202]: DEBUG nova.scheduler.client.report [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 831.328035] env[63202]: ERROR nova.compute.manager [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port cc6896d9-1f04-49c5-82f3-1d50d3ef5244, please check neutron logs for more information. [ 831.328035] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 831.328035] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 831.328035] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 831.328035] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 831.328035] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 831.328035] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 831.328035] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 831.328035] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 831.328035] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 831.328035] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 831.328035] env[63202]: ERROR nova.compute.manager raise self.value [ 831.328035] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 831.328035] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 831.328035] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 831.328035] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 831.328586] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 831.328586] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 831.328586] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port cc6896d9-1f04-49c5-82f3-1d50d3ef5244, please check neutron logs for more information. [ 831.328586] env[63202]: ERROR nova.compute.manager [ 831.328586] env[63202]: Traceback (most recent call last): [ 831.328586] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 831.328586] env[63202]: listener.cb(fileno) [ 831.328586] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 831.328586] env[63202]: result = function(*args, **kwargs) [ 831.328586] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 831.328586] env[63202]: return func(*args, **kwargs) [ 831.328586] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 831.328586] env[63202]: raise e [ 831.328586] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 831.328586] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 831.328586] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 831.328586] env[63202]: created_port_ids = self._update_ports_for_instance( [ 831.328586] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 831.328586] env[63202]: with excutils.save_and_reraise_exception(): [ 831.328586] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 831.328586] env[63202]: self.force_reraise() [ 831.328586] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 831.328586] env[63202]: raise self.value [ 831.328586] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 831.328586] env[63202]: updated_port = self._update_port( [ 831.328586] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 831.328586] env[63202]: _ensure_no_port_binding_failure(port) [ 831.328586] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 831.328586] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 831.330074] env[63202]: nova.exception.PortBindingFailed: Binding failed for port cc6896d9-1f04-49c5-82f3-1d50d3ef5244, please check neutron logs for more information. [ 831.330074] env[63202]: Removing descriptor: 16 [ 831.376517] env[63202]: DEBUG nova.compute.manager [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 831.401774] env[63202]: DEBUG nova.virt.hardware [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 831.401774] env[63202]: DEBUG nova.virt.hardware [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 831.401774] env[63202]: DEBUG nova.virt.hardware [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 831.402017] env[63202]: DEBUG nova.virt.hardware [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 831.402017] env[63202]: DEBUG nova.virt.hardware [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 831.402164] env[63202]: DEBUG nova.virt.hardware [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 831.402368] env[63202]: DEBUG nova.virt.hardware [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 831.402521] env[63202]: DEBUG nova.virt.hardware [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 831.402679] env[63202]: DEBUG nova.virt.hardware [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 831.402848] env[63202]: DEBUG nova.virt.hardware [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 831.403011] env[63202]: DEBUG nova.virt.hardware [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 831.403855] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e9fa5bb-de7a-4293-b95f-38f997e4d56e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.411983] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2c19ff9-0d11-4623-8067-b91ac26fd842 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.425768] env[63202]: ERROR nova.compute.manager [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port cc6896d9-1f04-49c5-82f3-1d50d3ef5244, please check neutron logs for more information. [ 831.425768] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] Traceback (most recent call last): [ 831.425768] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 831.425768] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] yield resources [ 831.425768] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 831.425768] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] self.driver.spawn(context, instance, image_meta, [ 831.425768] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 831.425768] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] self._vmops.spawn(context, instance, image_meta, injected_files, [ 831.425768] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 831.425768] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] vm_ref = self.build_virtual_machine(instance, [ 831.425768] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 831.426239] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] vif_infos = vmwarevif.get_vif_info(self._session, [ 831.426239] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 831.426239] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] for vif in network_info: [ 831.426239] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 831.426239] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] return self._sync_wrapper(fn, *args, **kwargs) [ 831.426239] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 831.426239] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] self.wait() [ 831.426239] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 831.426239] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] self[:] = self._gt.wait() [ 831.426239] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 831.426239] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] return self._exit_event.wait() [ 831.426239] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 831.426239] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] current.throw(*self._exc) [ 831.426692] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 831.426692] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] result = function(*args, **kwargs) [ 831.426692] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 831.426692] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] return func(*args, **kwargs) [ 831.426692] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 831.426692] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] raise e [ 831.426692] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 831.426692] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] nwinfo = self.network_api.allocate_for_instance( [ 831.426692] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 831.426692] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] created_port_ids = self._update_ports_for_instance( [ 831.426692] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 831.426692] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] with excutils.save_and_reraise_exception(): [ 831.426692] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 831.427280] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] self.force_reraise() [ 831.427280] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 831.427280] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] raise self.value [ 831.427280] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 831.427280] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] updated_port = self._update_port( [ 831.427280] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 831.427280] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] _ensure_no_port_binding_failure(port) [ 831.427280] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 831.427280] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] raise exception.PortBindingFailed(port_id=port['id']) [ 831.427280] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] nova.exception.PortBindingFailed: Binding failed for port cc6896d9-1f04-49c5-82f3-1d50d3ef5244, please check neutron logs for more information. [ 831.427280] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] [ 831.427280] env[63202]: INFO nova.compute.manager [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] Terminating instance [ 831.429497] env[63202]: DEBUG oslo_concurrency.lockutils [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Acquiring lock "refresh_cache-574ddcba-db96-4bd1-bb9e-3f8846c13430" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.429666] env[63202]: DEBUG oslo_concurrency.lockutils [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Acquired lock "refresh_cache-574ddcba-db96-4bd1-bb9e-3f8846c13430" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.429829] env[63202]: DEBUG nova.network.neutron [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 831.825961] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.473s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.827026] env[63202]: DEBUG nova.compute.manager [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 831.829497] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.046s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.842531] env[63202]: DEBUG nova.compute.manager [req-990c8c0f-95bb-45e0-8efd-44b29b7ba01b req-1a691836-a9bc-4aad-8fbd-164cd57ccd38 service nova] [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] Received event network-changed-cc6896d9-1f04-49c5-82f3-1d50d3ef5244 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 831.842531] env[63202]: DEBUG nova.compute.manager [req-990c8c0f-95bb-45e0-8efd-44b29b7ba01b req-1a691836-a9bc-4aad-8fbd-164cd57ccd38 service nova] [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] Refreshing instance network info cache due to event network-changed-cc6896d9-1f04-49c5-82f3-1d50d3ef5244. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 831.842531] env[63202]: DEBUG oslo_concurrency.lockutils [req-990c8c0f-95bb-45e0-8efd-44b29b7ba01b req-1a691836-a9bc-4aad-8fbd-164cd57ccd38 service nova] Acquiring lock "refresh_cache-574ddcba-db96-4bd1-bb9e-3f8846c13430" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.948102] env[63202]: DEBUG nova.network.neutron [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 832.036648] env[63202]: DEBUG nova.network.neutron [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.333816] env[63202]: DEBUG nova.compute.utils [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 832.338236] env[63202]: DEBUG nova.compute.manager [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 832.338460] env[63202]: DEBUG nova.network.neutron [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 832.506966] env[63202]: DEBUG nova.policy [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '393a67be05f348d0bf849d318adbffa5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2c62bc68629240f8abda8a25dc8c25ba', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 832.539528] env[63202]: DEBUG oslo_concurrency.lockutils [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Releasing lock "refresh_cache-574ddcba-db96-4bd1-bb9e-3f8846c13430" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.540367] env[63202]: DEBUG nova.compute.manager [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 832.540367] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 832.540367] env[63202]: DEBUG oslo_concurrency.lockutils [req-990c8c0f-95bb-45e0-8efd-44b29b7ba01b req-1a691836-a9bc-4aad-8fbd-164cd57ccd38 service nova] Acquired lock "refresh_cache-574ddcba-db96-4bd1-bb9e-3f8846c13430" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.540670] env[63202]: DEBUG nova.network.neutron [req-990c8c0f-95bb-45e0-8efd-44b29b7ba01b req-1a691836-a9bc-4aad-8fbd-164cd57ccd38 service nova] [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] Refreshing network info cache for port cc6896d9-1f04-49c5-82f3-1d50d3ef5244 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 832.541767] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6e94a126-0e7a-40aa-9554-8745f8064675 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.553222] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7036b5a8-a9c4-4822-b710-6d9ca4b02f69 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.577557] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 574ddcba-db96-4bd1-bb9e-3f8846c13430 could not be found. [ 832.577863] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 832.578123] env[63202]: INFO nova.compute.manager [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] Took 0.04 seconds to destroy the instance on the hypervisor. [ 832.578427] env[63202]: DEBUG oslo.service.loopingcall [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 832.580966] env[63202]: DEBUG nova.compute.manager [-] [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 832.581123] env[63202]: DEBUG nova.network.neutron [-] [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 832.602980] env[63202]: DEBUG nova.network.neutron [-] [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 832.607495] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93aef483-be8d-44da-91e7-b8ac57130da5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.615121] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb4bc650-56d3-421d-838e-ca85a72d3bd5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.648224] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edd026c5-45b9-4f1d-9cd6-21cffe7f0632 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.656071] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34173f6c-2fe0-44f2-8575-c929568ee9e9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.669874] env[63202]: DEBUG nova.compute.provider_tree [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 832.841502] env[63202]: DEBUG nova.compute.manager [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 832.881025] env[63202]: DEBUG nova.network.neutron [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] Successfully created port: 67590287-a9ba-4e46-93ad-f26ea015e2e6 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 833.062677] env[63202]: DEBUG nova.network.neutron [req-990c8c0f-95bb-45e0-8efd-44b29b7ba01b req-1a691836-a9bc-4aad-8fbd-164cd57ccd38 service nova] [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 833.105790] env[63202]: DEBUG nova.network.neutron [-] [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.145509] env[63202]: DEBUG nova.network.neutron [req-990c8c0f-95bb-45e0-8efd-44b29b7ba01b req-1a691836-a9bc-4aad-8fbd-164cd57ccd38 service nova] [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.172839] env[63202]: DEBUG nova.scheduler.client.report [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 833.610124] env[63202]: INFO nova.compute.manager [-] [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] Took 1.03 seconds to deallocate network for instance. [ 833.613057] env[63202]: DEBUG nova.compute.claims [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 833.613260] env[63202]: DEBUG oslo_concurrency.lockutils [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.648442] env[63202]: DEBUG oslo_concurrency.lockutils [req-990c8c0f-95bb-45e0-8efd-44b29b7ba01b req-1a691836-a9bc-4aad-8fbd-164cd57ccd38 service nova] Releasing lock "refresh_cache-574ddcba-db96-4bd1-bb9e-3f8846c13430" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.648697] env[63202]: DEBUG nova.compute.manager [req-990c8c0f-95bb-45e0-8efd-44b29b7ba01b req-1a691836-a9bc-4aad-8fbd-164cd57ccd38 service nova] [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] Received event network-vif-deleted-cc6896d9-1f04-49c5-82f3-1d50d3ef5244 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 833.678338] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.849s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.678931] env[63202]: ERROR nova.compute.manager [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ccc1a7cc-c159-499e-9bf1-76d48a9127b3, please check neutron logs for more information. [ 833.678931] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] Traceback (most recent call last): [ 833.678931] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 833.678931] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] self.driver.spawn(context, instance, image_meta, [ 833.678931] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 833.678931] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 833.678931] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 833.678931] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] vm_ref = self.build_virtual_machine(instance, [ 833.678931] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 833.678931] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] vif_infos = vmwarevif.get_vif_info(self._session, [ 833.678931] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 833.679277] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] for vif in network_info: [ 833.679277] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 833.679277] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] return self._sync_wrapper(fn, *args, **kwargs) [ 833.679277] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 833.679277] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] self.wait() [ 833.679277] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 833.679277] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] self[:] = self._gt.wait() [ 833.679277] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 833.679277] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] return self._exit_event.wait() [ 833.679277] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 833.679277] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] current.throw(*self._exc) [ 833.679277] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 833.679277] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] result = function(*args, **kwargs) [ 833.679716] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 833.679716] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] return func(*args, **kwargs) [ 833.679716] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 833.679716] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] raise e [ 833.679716] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 833.679716] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] nwinfo = self.network_api.allocate_for_instance( [ 833.679716] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 833.679716] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] created_port_ids = self._update_ports_for_instance( [ 833.679716] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 833.679716] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] with excutils.save_and_reraise_exception(): [ 833.679716] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 833.679716] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] self.force_reraise() [ 833.679716] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 833.680083] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] raise self.value [ 833.680083] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 833.680083] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] updated_port = self._update_port( [ 833.680083] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 833.680083] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] _ensure_no_port_binding_failure(port) [ 833.680083] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 833.680083] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] raise exception.PortBindingFailed(port_id=port['id']) [ 833.680083] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] nova.exception.PortBindingFailed: Binding failed for port ccc1a7cc-c159-499e-9bf1-76d48a9127b3, please check neutron logs for more information. [ 833.680083] env[63202]: ERROR nova.compute.manager [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] [ 833.680083] env[63202]: DEBUG nova.compute.utils [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] Binding failed for port ccc1a7cc-c159-499e-9bf1-76d48a9127b3, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 833.680842] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.758s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.683921] env[63202]: DEBUG nova.compute.manager [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] Build of instance e24b7994-3786-4709-bcb5-e0b2c9731f2a was re-scheduled: Binding failed for port ccc1a7cc-c159-499e-9bf1-76d48a9127b3, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 833.684350] env[63202]: DEBUG nova.compute.manager [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 833.684591] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Acquiring lock "refresh_cache-e24b7994-3786-4709-bcb5-e0b2c9731f2a" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.684707] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Acquired lock "refresh_cache-e24b7994-3786-4709-bcb5-e0b2c9731f2a" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.684850] env[63202]: DEBUG nova.network.neutron [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 833.856018] env[63202]: DEBUG nova.compute.manager [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 833.871590] env[63202]: DEBUG nova.compute.manager [req-675a16d6-ea3e-4135-b414-bbfc0ba2c2fc req-12ea99b1-fb0f-4455-9e48-c5db04021b69 service nova] [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] Received event network-changed-67590287-a9ba-4e46-93ad-f26ea015e2e6 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 833.871759] env[63202]: DEBUG nova.compute.manager [req-675a16d6-ea3e-4135-b414-bbfc0ba2c2fc req-12ea99b1-fb0f-4455-9e48-c5db04021b69 service nova] [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] Refreshing instance network info cache due to event network-changed-67590287-a9ba-4e46-93ad-f26ea015e2e6. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 833.871972] env[63202]: DEBUG oslo_concurrency.lockutils [req-675a16d6-ea3e-4135-b414-bbfc0ba2c2fc req-12ea99b1-fb0f-4455-9e48-c5db04021b69 service nova] Acquiring lock "refresh_cache-57f446ee-edf3-40e7-96bf-8fdce62bb19f" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.872137] env[63202]: DEBUG oslo_concurrency.lockutils [req-675a16d6-ea3e-4135-b414-bbfc0ba2c2fc req-12ea99b1-fb0f-4455-9e48-c5db04021b69 service nova] Acquired lock "refresh_cache-57f446ee-edf3-40e7-96bf-8fdce62bb19f" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.872283] env[63202]: DEBUG nova.network.neutron [req-675a16d6-ea3e-4135-b414-bbfc0ba2c2fc req-12ea99b1-fb0f-4455-9e48-c5db04021b69 service nova] [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] Refreshing network info cache for port 67590287-a9ba-4e46-93ad-f26ea015e2e6 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 833.887409] env[63202]: ERROR nova.compute.manager [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 67590287-a9ba-4e46-93ad-f26ea015e2e6, please check neutron logs for more information. [ 833.887409] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 833.887409] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 833.887409] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 833.887409] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 833.887409] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 833.887409] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 833.887409] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 833.887409] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 833.887409] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 833.887409] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 833.887409] env[63202]: ERROR nova.compute.manager raise self.value [ 833.887409] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 833.887409] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 833.887409] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 833.887409] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 833.888265] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 833.888265] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 833.888265] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 67590287-a9ba-4e46-93ad-f26ea015e2e6, please check neutron logs for more information. [ 833.888265] env[63202]: ERROR nova.compute.manager [ 833.888265] env[63202]: Traceback (most recent call last): [ 833.888265] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 833.888265] env[63202]: listener.cb(fileno) [ 833.888265] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 833.888265] env[63202]: result = function(*args, **kwargs) [ 833.888265] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 833.888265] env[63202]: return func(*args, **kwargs) [ 833.888265] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 833.888265] env[63202]: raise e [ 833.888265] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 833.888265] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 833.888265] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 833.888265] env[63202]: created_port_ids = self._update_ports_for_instance( [ 833.888265] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 833.888265] env[63202]: with excutils.save_and_reraise_exception(): [ 833.888265] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 833.888265] env[63202]: self.force_reraise() [ 833.888265] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 833.888265] env[63202]: raise self.value [ 833.888265] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 833.888265] env[63202]: updated_port = self._update_port( [ 833.888265] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 833.888265] env[63202]: _ensure_no_port_binding_failure(port) [ 833.888265] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 833.888265] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 833.889554] env[63202]: nova.exception.PortBindingFailed: Binding failed for port 67590287-a9ba-4e46-93ad-f26ea015e2e6, please check neutron logs for more information. [ 833.889554] env[63202]: Removing descriptor: 16 [ 833.889554] env[63202]: DEBUG nova.virt.hardware [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 833.889872] env[63202]: DEBUG nova.virt.hardware [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 833.889872] env[63202]: DEBUG nova.virt.hardware [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 833.890090] env[63202]: DEBUG nova.virt.hardware [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 833.891015] env[63202]: DEBUG nova.virt.hardware [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 833.891215] env[63202]: DEBUG nova.virt.hardware [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 833.891451] env[63202]: DEBUG nova.virt.hardware [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 833.891616] env[63202]: DEBUG nova.virt.hardware [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 833.891819] env[63202]: DEBUG nova.virt.hardware [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 833.892156] env[63202]: DEBUG nova.virt.hardware [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 833.892254] env[63202]: DEBUG nova.virt.hardware [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 833.893474] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b13ee6f4-c2fe-46ed-8703-d79ff939a40c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.902594] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-345606d7-9ae7-4da4-92c8-a80ce29b6f71 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.916801] env[63202]: ERROR nova.compute.manager [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 67590287-a9ba-4e46-93ad-f26ea015e2e6, please check neutron logs for more information. [ 833.916801] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] Traceback (most recent call last): [ 833.916801] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 833.916801] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] yield resources [ 833.916801] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 833.916801] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] self.driver.spawn(context, instance, image_meta, [ 833.916801] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 833.916801] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 833.916801] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 833.916801] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] vm_ref = self.build_virtual_machine(instance, [ 833.916801] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 833.917536] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] vif_infos = vmwarevif.get_vif_info(self._session, [ 833.917536] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 833.917536] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] for vif in network_info: [ 833.917536] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 833.917536] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] return self._sync_wrapper(fn, *args, **kwargs) [ 833.917536] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 833.917536] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] self.wait() [ 833.917536] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 833.917536] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] self[:] = self._gt.wait() [ 833.917536] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 833.917536] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] return self._exit_event.wait() [ 833.917536] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 833.917536] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] current.throw(*self._exc) [ 833.918056] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 833.918056] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] result = function(*args, **kwargs) [ 833.918056] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 833.918056] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] return func(*args, **kwargs) [ 833.918056] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 833.918056] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] raise e [ 833.918056] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 833.918056] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] nwinfo = self.network_api.allocate_for_instance( [ 833.918056] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 833.918056] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] created_port_ids = self._update_ports_for_instance( [ 833.918056] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 833.918056] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] with excutils.save_and_reraise_exception(): [ 833.918056] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 833.918421] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] self.force_reraise() [ 833.918421] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 833.918421] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] raise self.value [ 833.918421] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 833.918421] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] updated_port = self._update_port( [ 833.918421] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 833.918421] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] _ensure_no_port_binding_failure(port) [ 833.918421] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 833.918421] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] raise exception.PortBindingFailed(port_id=port['id']) [ 833.918421] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] nova.exception.PortBindingFailed: Binding failed for port 67590287-a9ba-4e46-93ad-f26ea015e2e6, please check neutron logs for more information. [ 833.918421] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] [ 833.918421] env[63202]: INFO nova.compute.manager [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] Terminating instance [ 833.919491] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] Acquiring lock "refresh_cache-57f446ee-edf3-40e7-96bf-8fdce62bb19f" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.202027] env[63202]: DEBUG nova.network.neutron [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 834.293085] env[63202]: DEBUG nova.network.neutron [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.392012] env[63202]: DEBUG nova.network.neutron [req-675a16d6-ea3e-4135-b414-bbfc0ba2c2fc req-12ea99b1-fb0f-4455-9e48-c5db04021b69 service nova] [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 834.402212] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-333df0a8-aaf2-43b1-8700-715bb1afef9c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.409903] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-166a3227-742f-4703-ac2b-71e31054d54b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.444285] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4d0983a-6aaf-48dd-a649-4e0fa5b0c355 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.451749] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c871f6d0-adc9-4d5b-8f55-2522a7f91b32 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.465212] env[63202]: DEBUG nova.compute.provider_tree [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 834.482547] env[63202]: DEBUG nova.network.neutron [req-675a16d6-ea3e-4135-b414-bbfc0ba2c2fc req-12ea99b1-fb0f-4455-9e48-c5db04021b69 service nova] [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.795771] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Releasing lock "refresh_cache-e24b7994-3786-4709-bcb5-e0b2c9731f2a" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.796077] env[63202]: DEBUG nova.compute.manager [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 834.796201] env[63202]: DEBUG nova.compute.manager [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 834.796455] env[63202]: DEBUG nova.network.neutron [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 834.816017] env[63202]: DEBUG nova.network.neutron [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 834.968791] env[63202]: DEBUG nova.scheduler.client.report [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 834.985773] env[63202]: DEBUG oslo_concurrency.lockutils [req-675a16d6-ea3e-4135-b414-bbfc0ba2c2fc req-12ea99b1-fb0f-4455-9e48-c5db04021b69 service nova] Releasing lock "refresh_cache-57f446ee-edf3-40e7-96bf-8fdce62bb19f" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.986723] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] Acquired lock "refresh_cache-57f446ee-edf3-40e7-96bf-8fdce62bb19f" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.987040] env[63202]: DEBUG nova.network.neutron [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 835.318372] env[63202]: DEBUG nova.network.neutron [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.474991] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.794s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.475684] env[63202]: ERROR nova.compute.manager [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 965fe5c5-bcdd-42fb-912d-ad9870e37bf0, please check neutron logs for more information. [ 835.475684] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] Traceback (most recent call last): [ 835.475684] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 835.475684] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] self.driver.spawn(context, instance, image_meta, [ 835.475684] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 835.475684] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 835.475684] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 835.475684] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] vm_ref = self.build_virtual_machine(instance, [ 835.475684] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 835.475684] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] vif_infos = vmwarevif.get_vif_info(self._session, [ 835.475684] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 835.476062] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] for vif in network_info: [ 835.476062] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 835.476062] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] return self._sync_wrapper(fn, *args, **kwargs) [ 835.476062] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 835.476062] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] self.wait() [ 835.476062] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 835.476062] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] self[:] = self._gt.wait() [ 835.476062] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 835.476062] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] return self._exit_event.wait() [ 835.476062] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 835.476062] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] current.throw(*self._exc) [ 835.476062] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 835.476062] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] result = function(*args, **kwargs) [ 835.476468] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 835.476468] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] return func(*args, **kwargs) [ 835.476468] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 835.476468] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] raise e [ 835.476468] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 835.476468] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] nwinfo = self.network_api.allocate_for_instance( [ 835.476468] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 835.476468] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] created_port_ids = self._update_ports_for_instance( [ 835.476468] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 835.476468] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] with excutils.save_and_reraise_exception(): [ 835.476468] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 835.476468] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] self.force_reraise() [ 835.476468] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 835.476881] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] raise self.value [ 835.476881] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 835.476881] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] updated_port = self._update_port( [ 835.476881] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 835.476881] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] _ensure_no_port_binding_failure(port) [ 835.476881] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 835.476881] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] raise exception.PortBindingFailed(port_id=port['id']) [ 835.476881] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] nova.exception.PortBindingFailed: Binding failed for port 965fe5c5-bcdd-42fb-912d-ad9870e37bf0, please check neutron logs for more information. [ 835.476881] env[63202]: ERROR nova.compute.manager [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] [ 835.476881] env[63202]: DEBUG nova.compute.utils [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] Binding failed for port 965fe5c5-bcdd-42fb-912d-ad9870e37bf0, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 835.477933] env[63202]: DEBUG oslo_concurrency.lockutils [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.317s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.481297] env[63202]: DEBUG nova.compute.manager [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] Build of instance d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6 was re-scheduled: Binding failed for port 965fe5c5-bcdd-42fb-912d-ad9870e37bf0, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 835.481738] env[63202]: DEBUG nova.compute.manager [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 835.481956] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] Acquiring lock "refresh_cache-d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.482111] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] Acquired lock "refresh_cache-d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.482268] env[63202]: DEBUG nova.network.neutron [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 835.509610] env[63202]: DEBUG nova.network.neutron [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 835.584686] env[63202]: DEBUG nova.network.neutron [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.821509] env[63202]: INFO nova.compute.manager [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] [instance: e24b7994-3786-4709-bcb5-e0b2c9731f2a] Took 1.02 seconds to deallocate network for instance. [ 835.899470] env[63202]: DEBUG nova.compute.manager [req-1de17486-b92e-4b3a-acd6-292782617780 req-fdbc1f87-41b7-46ca-8242-f8bd3fd81775 service nova] [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] Received event network-vif-deleted-67590287-a9ba-4e46-93ad-f26ea015e2e6 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 835.999305] env[63202]: DEBUG nova.network.neutron [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 836.083147] env[63202]: DEBUG nova.network.neutron [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.088429] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] Releasing lock "refresh_cache-57f446ee-edf3-40e7-96bf-8fdce62bb19f" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.088816] env[63202]: DEBUG nova.compute.manager [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 836.089008] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 836.090086] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-eee25d9c-cd05-4664-ab82-156caf90611a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.098583] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d109e39-dfc0-4cca-b2a7-e27ce4797853 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.122710] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 57f446ee-edf3-40e7-96bf-8fdce62bb19f could not be found. [ 836.122935] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 836.123129] env[63202]: INFO nova.compute.manager [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] Took 0.03 seconds to destroy the instance on the hypervisor. [ 836.123390] env[63202]: DEBUG oslo.service.loopingcall [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 836.125778] env[63202]: DEBUG nova.compute.manager [-] [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 836.125881] env[63202]: DEBUG nova.network.neutron [-] [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 836.145079] env[63202]: DEBUG nova.network.neutron [-] [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 836.239933] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2db1bb07-7199-4a96-b758-8f15d66fde24 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.247283] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ca5e75b-d418-4fae-b179-5b2e8595275c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.276850] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-696e9519-e28a-46ac-93b7-a706164cce60 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.283754] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb8c9b87-3740-4353-9bf6-412e9a7b7a06 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.298119] env[63202]: DEBUG nova.compute.provider_tree [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 836.585843] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] Releasing lock "refresh_cache-d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.586103] env[63202]: DEBUG nova.compute.manager [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 836.586282] env[63202]: DEBUG nova.compute.manager [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 836.586445] env[63202]: DEBUG nova.network.neutron [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 836.601882] env[63202]: DEBUG nova.network.neutron [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 836.649757] env[63202]: DEBUG nova.network.neutron [-] [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.801474] env[63202]: DEBUG nova.scheduler.client.report [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 836.849722] env[63202]: INFO nova.scheduler.client.report [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Deleted allocations for instance e24b7994-3786-4709-bcb5-e0b2c9731f2a [ 837.105030] env[63202]: DEBUG nova.network.neutron [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.152797] env[63202]: INFO nova.compute.manager [-] [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] Took 1.03 seconds to deallocate network for instance. [ 837.155144] env[63202]: DEBUG nova.compute.claims [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 837.155324] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.306723] env[63202]: DEBUG oslo_concurrency.lockutils [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.829s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.307394] env[63202]: ERROR nova.compute.manager [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 560f6e99-7a69-440c-bee5-e88a343edd85, please check neutron logs for more information. [ 837.307394] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] Traceback (most recent call last): [ 837.307394] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 837.307394] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] self.driver.spawn(context, instance, image_meta, [ 837.307394] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 837.307394] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 837.307394] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 837.307394] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] vm_ref = self.build_virtual_machine(instance, [ 837.307394] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 837.307394] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] vif_infos = vmwarevif.get_vif_info(self._session, [ 837.307394] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 837.307855] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] for vif in network_info: [ 837.307855] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 837.307855] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] return self._sync_wrapper(fn, *args, **kwargs) [ 837.307855] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 837.307855] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] self.wait() [ 837.307855] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 837.307855] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] self[:] = self._gt.wait() [ 837.307855] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 837.307855] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] return self._exit_event.wait() [ 837.307855] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 837.307855] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] current.throw(*self._exc) [ 837.307855] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 837.307855] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] result = function(*args, **kwargs) [ 837.308300] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 837.308300] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] return func(*args, **kwargs) [ 837.308300] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 837.308300] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] raise e [ 837.308300] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 837.308300] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] nwinfo = self.network_api.allocate_for_instance( [ 837.308300] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 837.308300] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] created_port_ids = self._update_ports_for_instance( [ 837.308300] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 837.308300] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] with excutils.save_and_reraise_exception(): [ 837.308300] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 837.308300] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] self.force_reraise() [ 837.308300] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 837.308790] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] raise self.value [ 837.308790] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 837.308790] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] updated_port = self._update_port( [ 837.308790] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 837.308790] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] _ensure_no_port_binding_failure(port) [ 837.308790] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 837.308790] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] raise exception.PortBindingFailed(port_id=port['id']) [ 837.308790] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] nova.exception.PortBindingFailed: Binding failed for port 560f6e99-7a69-440c-bee5-e88a343edd85, please check neutron logs for more information. [ 837.308790] env[63202]: ERROR nova.compute.manager [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] [ 837.308790] env[63202]: DEBUG nova.compute.utils [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] Binding failed for port 560f6e99-7a69-440c-bee5-e88a343edd85, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 837.310944] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.586s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.312635] env[63202]: DEBUG nova.compute.manager [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] Build of instance 95d2a2d7-443b-43d6-97ff-fe36fd11002f was re-scheduled: Binding failed for port 560f6e99-7a69-440c-bee5-e88a343edd85, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 837.313050] env[63202]: DEBUG nova.compute.manager [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 837.313279] env[63202]: DEBUG oslo_concurrency.lockutils [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] Acquiring lock "refresh_cache-95d2a2d7-443b-43d6-97ff-fe36fd11002f" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.313423] env[63202]: DEBUG oslo_concurrency.lockutils [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] Acquired lock "refresh_cache-95d2a2d7-443b-43d6-97ff-fe36fd11002f" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.313576] env[63202]: DEBUG nova.network.neutron [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 837.360293] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c2dd0727-1815-4fdd-ae68-6ae4d6483abd tempest-ServersAdminTestJSON-1081365632 tempest-ServersAdminTestJSON-1081365632-project-member] Lock "e24b7994-3786-4709-bcb5-e0b2c9731f2a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 184.379s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.606691] env[63202]: INFO nova.compute.manager [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] [instance: d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6] Took 1.02 seconds to deallocate network for instance. [ 837.854807] env[63202]: DEBUG nova.network.neutron [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 837.862702] env[63202]: DEBUG nova.compute.manager [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 838.046830] env[63202]: DEBUG nova.network.neutron [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.072645] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7350462e-caf3-4336-a626-a2ec087143c0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.080757] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52c7ad1d-73a6-4de6-8a2c-4ea776cf8821 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.113902] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-613d11d2-d0e0-4747-9ade-8e2c8f983ea7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.122487] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab13db27-172f-4b30-9d42-c702ffadc651 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.142108] env[63202]: DEBUG nova.compute.provider_tree [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 838.392906] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.549222] env[63202]: DEBUG oslo_concurrency.lockutils [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] Releasing lock "refresh_cache-95d2a2d7-443b-43d6-97ff-fe36fd11002f" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.550123] env[63202]: DEBUG nova.compute.manager [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 838.550424] env[63202]: DEBUG nova.compute.manager [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 838.552627] env[63202]: DEBUG nova.network.neutron [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 838.567770] env[63202]: DEBUG nova.network.neutron [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 838.638106] env[63202]: INFO nova.scheduler.client.report [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] Deleted allocations for instance d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6 [ 838.648822] env[63202]: DEBUG nova.scheduler.client.report [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 839.073207] env[63202]: DEBUG nova.network.neutron [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.148489] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2fef9c13-354f-4963-b0b4-9bc739ce579a tempest-InstanceActionsTestJSON-2048493098 tempest-InstanceActionsTestJSON-2048493098-project-member] Lock "d71bc7f9-29d4-43d6-a17a-80ae6f4ec4a6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 179.771s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.154062] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.844s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.154062] env[63202]: ERROR nova.compute.manager [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a9945f05-55e7-4534-a38e-d921a091bb11, please check neutron logs for more information. [ 839.154062] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] Traceback (most recent call last): [ 839.154062] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 839.154062] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] self.driver.spawn(context, instance, image_meta, [ 839.154062] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 839.154062] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 839.154062] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 839.154062] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] vm_ref = self.build_virtual_machine(instance, [ 839.154438] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 839.154438] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] vif_infos = vmwarevif.get_vif_info(self._session, [ 839.154438] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 839.154438] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] for vif in network_info: [ 839.154438] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 839.154438] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] return self._sync_wrapper(fn, *args, **kwargs) [ 839.154438] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 839.154438] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] self.wait() [ 839.154438] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 839.154438] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] self[:] = self._gt.wait() [ 839.154438] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 839.154438] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] return self._exit_event.wait() [ 839.154438] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 839.154800] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] result = hub.switch() [ 839.154800] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 839.154800] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] return self.greenlet.switch() [ 839.154800] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 839.154800] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] result = function(*args, **kwargs) [ 839.154800] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 839.154800] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] return func(*args, **kwargs) [ 839.154800] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 839.154800] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] raise e [ 839.154800] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 839.154800] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] nwinfo = self.network_api.allocate_for_instance( [ 839.154800] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 839.154800] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] created_port_ids = self._update_ports_for_instance( [ 839.155060] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 839.155060] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] with excutils.save_and_reraise_exception(): [ 839.155060] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 839.155060] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] self.force_reraise() [ 839.155060] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 839.155060] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] raise self.value [ 839.155060] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 839.155060] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] updated_port = self._update_port( [ 839.155060] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 839.155060] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] _ensure_no_port_binding_failure(port) [ 839.155060] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 839.155060] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] raise exception.PortBindingFailed(port_id=port['id']) [ 839.155373] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] nova.exception.PortBindingFailed: Binding failed for port a9945f05-55e7-4534-a38e-d921a091bb11, please check neutron logs for more information. [ 839.155373] env[63202]: ERROR nova.compute.manager [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] [ 839.155373] env[63202]: DEBUG nova.compute.utils [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] Binding failed for port a9945f05-55e7-4534-a38e-d921a091bb11, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 839.157244] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 15.360s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.157244] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.157244] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63202) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 839.157505] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.201s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.160944] env[63202]: DEBUG nova.compute.manager [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] Build of instance 39d9b91b-c009-40e1-ad2b-ebef650188a4 was re-scheduled: Binding failed for port a9945f05-55e7-4534-a38e-d921a091bb11, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 839.161523] env[63202]: DEBUG nova.compute.manager [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 839.161953] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquiring lock "refresh_cache-39d9b91b-c009-40e1-ad2b-ebef650188a4" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.161953] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquired lock "refresh_cache-39d9b91b-c009-40e1-ad2b-ebef650188a4" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.162640] env[63202]: DEBUG nova.network.neutron [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 839.164146] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd0cb02f-a09d-49db-9989-dda2f33811c0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.177510] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0199270a-5d1c-4ebb-8872-48b6ecb8fb67 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.195257] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba588307-0639-486a-a249-038acc117821 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.207536] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2fa63e3-c3dd-4185-b612-b9898dd3318a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.239836] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181350MB free_disk=176GB free_vcpus=48 pci_devices=None {{(pid=63202) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 839.239836] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.581022] env[63202]: INFO nova.compute.manager [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] [instance: 95d2a2d7-443b-43d6-97ff-fe36fd11002f] Took 1.03 seconds to deallocate network for instance. [ 839.650585] env[63202]: DEBUG nova.compute.manager [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 839.699669] env[63202]: DEBUG nova.network.neutron [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 839.827045] env[63202]: DEBUG nova.network.neutron [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.933080] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36065e41-4408-445f-98a7-66f622975248 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.941572] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50e4354d-fe19-4ab3-861e-8edd5e86a3e9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.974279] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7b839da-7b80-4371-b962-205a0d3b9390 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.981965] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8176979c-d3c9-4f98-92ee-028d40b1af2b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.000680] env[63202]: DEBUG nova.compute.provider_tree [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 840.169570] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.329721] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Releasing lock "refresh_cache-39d9b91b-c009-40e1-ad2b-ebef650188a4" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.332211] env[63202]: DEBUG nova.compute.manager [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 840.332211] env[63202]: DEBUG nova.compute.manager [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 840.332211] env[63202]: DEBUG nova.network.neutron [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 840.349012] env[63202]: DEBUG nova.network.neutron [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 840.503652] env[63202]: DEBUG nova.scheduler.client.report [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 840.630632] env[63202]: INFO nova.scheduler.client.report [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] Deleted allocations for instance 95d2a2d7-443b-43d6-97ff-fe36fd11002f [ 840.851411] env[63202]: DEBUG nova.network.neutron [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.008736] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.851s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.009104] env[63202]: ERROR nova.compute.manager [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port eddbfe5e-fc77-4e94-9795-fb3364bdbd51, please check neutron logs for more information. [ 841.009104] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] Traceback (most recent call last): [ 841.009104] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 841.009104] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] self.driver.spawn(context, instance, image_meta, [ 841.009104] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 841.009104] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 841.009104] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 841.009104] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] vm_ref = self.build_virtual_machine(instance, [ 841.009104] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 841.009104] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] vif_infos = vmwarevif.get_vif_info(self._session, [ 841.009104] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 841.009379] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] for vif in network_info: [ 841.009379] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 841.009379] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] return self._sync_wrapper(fn, *args, **kwargs) [ 841.009379] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 841.009379] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] self.wait() [ 841.009379] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 841.009379] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] self[:] = self._gt.wait() [ 841.009379] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 841.009379] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] return self._exit_event.wait() [ 841.009379] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 841.009379] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] current.throw(*self._exc) [ 841.009379] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 841.009379] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] result = function(*args, **kwargs) [ 841.009643] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 841.009643] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] return func(*args, **kwargs) [ 841.009643] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 841.009643] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] raise e [ 841.009643] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 841.009643] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] nwinfo = self.network_api.allocate_for_instance( [ 841.009643] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 841.009643] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] created_port_ids = self._update_ports_for_instance( [ 841.009643] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 841.009643] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] with excutils.save_and_reraise_exception(): [ 841.009643] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 841.009643] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] self.force_reraise() [ 841.009643] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 841.010214] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] raise self.value [ 841.010214] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 841.010214] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] updated_port = self._update_port( [ 841.010214] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 841.010214] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] _ensure_no_port_binding_failure(port) [ 841.010214] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 841.010214] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] raise exception.PortBindingFailed(port_id=port['id']) [ 841.010214] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] nova.exception.PortBindingFailed: Binding failed for port eddbfe5e-fc77-4e94-9795-fb3364bdbd51, please check neutron logs for more information. [ 841.010214] env[63202]: ERROR nova.compute.manager [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] [ 841.010214] env[63202]: DEBUG nova.compute.utils [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] Binding failed for port eddbfe5e-fc77-4e94-9795-fb3364bdbd51, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 841.011505] env[63202]: DEBUG oslo_concurrency.lockutils [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.585s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.015277] env[63202]: DEBUG nova.compute.manager [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] Build of instance 06bd8147-5479-4c5e-8ba2-6d3a205ba05e was re-scheduled: Binding failed for port eddbfe5e-fc77-4e94-9795-fb3364bdbd51, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 841.015816] env[63202]: DEBUG nova.compute.manager [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 841.016119] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Acquiring lock "refresh_cache-06bd8147-5479-4c5e-8ba2-6d3a205ba05e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.016262] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Acquired lock "refresh_cache-06bd8147-5479-4c5e-8ba2-6d3a205ba05e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.016450] env[63202]: DEBUG nova.network.neutron [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 841.139155] env[63202]: DEBUG oslo_concurrency.lockutils [None req-539faead-409c-4603-8579-2183bdc0ca10 tempest-ServerTagsTestJSON-719409431 tempest-ServerTagsTestJSON-719409431-project-member] Lock "95d2a2d7-443b-43d6-97ff-fe36fd11002f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 177.990s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.354901] env[63202]: INFO nova.compute.manager [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 39d9b91b-c009-40e1-ad2b-ebef650188a4] Took 1.02 seconds to deallocate network for instance. [ 841.539925] env[63202]: DEBUG nova.network.neutron [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 841.644074] env[63202]: DEBUG nova.compute.manager [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 841.711994] env[63202]: DEBUG nova.network.neutron [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.773466] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d195826-6fc3-4423-9665-6dade457102a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.784718] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7688e43a-5143-47a8-ad6a-19e5738b1438 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.815303] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa5c75d0-c694-4cf5-aa8f-180c98262e68 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.822693] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b471e44-a7bc-490d-8c00-08d2c547aed3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.836939] env[63202]: DEBUG nova.compute.provider_tree [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 842.163361] env[63202]: DEBUG oslo_concurrency.lockutils [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.217390] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Releasing lock "refresh_cache-06bd8147-5479-4c5e-8ba2-6d3a205ba05e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.217728] env[63202]: DEBUG nova.compute.manager [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 842.218019] env[63202]: DEBUG nova.compute.manager [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 842.218783] env[63202]: DEBUG nova.network.neutron [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 842.319854] env[63202]: DEBUG nova.network.neutron [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 842.340498] env[63202]: DEBUG nova.scheduler.client.report [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 842.396634] env[63202]: INFO nova.scheduler.client.report [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Deleted allocations for instance 39d9b91b-c009-40e1-ad2b-ebef650188a4 [ 842.823035] env[63202]: DEBUG nova.network.neutron [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.845887] env[63202]: DEBUG oslo_concurrency.lockutils [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.835s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.847106] env[63202]: ERROR nova.compute.manager [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a89d882c-6b78-4f47-929d-8d342776d561, please check neutron logs for more information. [ 842.847106] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] Traceback (most recent call last): [ 842.847106] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 842.847106] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] self.driver.spawn(context, instance, image_meta, [ 842.847106] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 842.847106] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 842.847106] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 842.847106] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] vm_ref = self.build_virtual_machine(instance, [ 842.847106] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 842.847106] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] vif_infos = vmwarevif.get_vif_info(self._session, [ 842.847106] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 842.847502] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] for vif in network_info: [ 842.847502] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 842.847502] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] return self._sync_wrapper(fn, *args, **kwargs) [ 842.847502] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 842.847502] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] self.wait() [ 842.847502] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 842.847502] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] self[:] = self._gt.wait() [ 842.847502] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 842.847502] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] return self._exit_event.wait() [ 842.847502] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 842.847502] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] result = hub.switch() [ 842.847502] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 842.847502] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] return self.greenlet.switch() [ 842.847906] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 842.847906] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] result = function(*args, **kwargs) [ 842.847906] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 842.847906] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] return func(*args, **kwargs) [ 842.847906] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 842.847906] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] raise e [ 842.847906] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 842.847906] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] nwinfo = self.network_api.allocate_for_instance( [ 842.847906] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 842.847906] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] created_port_ids = self._update_ports_for_instance( [ 842.847906] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 842.847906] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] with excutils.save_and_reraise_exception(): [ 842.847906] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 842.848244] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] self.force_reraise() [ 842.848244] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 842.848244] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] raise self.value [ 842.848244] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 842.848244] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] updated_port = self._update_port( [ 842.848244] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 842.848244] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] _ensure_no_port_binding_failure(port) [ 842.848244] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 842.848244] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] raise exception.PortBindingFailed(port_id=port['id']) [ 842.848244] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] nova.exception.PortBindingFailed: Binding failed for port a89d882c-6b78-4f47-929d-8d342776d561, please check neutron logs for more information. [ 842.848244] env[63202]: ERROR nova.compute.manager [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] [ 842.848525] env[63202]: DEBUG nova.compute.utils [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] Binding failed for port a89d882c-6b78-4f47-929d-8d342776d561, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 842.849051] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.021s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.852311] env[63202]: DEBUG nova.compute.manager [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] Build of instance 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a was re-scheduled: Binding failed for port a89d882c-6b78-4f47-929d-8d342776d561, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 842.852518] env[63202]: DEBUG nova.compute.manager [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 842.852775] env[63202]: DEBUG oslo_concurrency.lockutils [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] Acquiring lock "refresh_cache-56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.852946] env[63202]: DEBUG oslo_concurrency.lockutils [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] Acquired lock "refresh_cache-56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.853128] env[63202]: DEBUG nova.network.neutron [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 842.908300] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f5ce7826-4e5d-4cd2-bd34-3f6897019727 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lock "39d9b91b-c009-40e1-ad2b-ebef650188a4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 173.152s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.328410] env[63202]: INFO nova.compute.manager [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] [instance: 06bd8147-5479-4c5e-8ba2-6d3a205ba05e] Took 1.11 seconds to deallocate network for instance. [ 843.333560] env[63202]: DEBUG oslo_concurrency.lockutils [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Acquiring lock "36291628-af7d-43cf-8149-09f57df47890" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.333789] env[63202]: DEBUG oslo_concurrency.lockutils [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Lock "36291628-af7d-43cf-8149-09f57df47890" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.390211] env[63202]: DEBUG nova.network.neutron [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 843.411718] env[63202]: DEBUG nova.compute.manager [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: d674db7a-8cf0-46f8-b676-fced0767c105] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 843.536377] env[63202]: DEBUG nova.network.neutron [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.599319] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7eba749-b049-4a3d-bb26-a432c44946a0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.607221] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed150222-4a55-41c2-926c-21807afe9bf4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.637859] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a74182b5-db18-450c-a7c0-cf5851d84d0f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.643494] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Acquiring lock "3d1518d3-bed2-4b7b-af1d-d7da49e92874" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.643901] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Lock "3d1518d3-bed2-4b7b-af1d-d7da49e92874" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.648945] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddd76ef6-9892-4133-a1d0-89eae34d0558 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.665016] env[63202]: DEBUG nova.compute.provider_tree [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 843.944214] env[63202]: DEBUG oslo_concurrency.lockutils [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.959267] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Acquiring lock "61a6528e-4e4e-49b7-be7b-dd30bd8023d8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.959488] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Lock "61a6528e-4e4e-49b7-be7b-dd30bd8023d8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.042758] env[63202]: DEBUG oslo_concurrency.lockutils [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] Releasing lock "refresh_cache-56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.042992] env[63202]: DEBUG nova.compute.manager [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 844.043172] env[63202]: DEBUG nova.compute.manager [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 844.043339] env[63202]: DEBUG nova.network.neutron [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 844.061123] env[63202]: DEBUG nova.network.neutron [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 844.171461] env[63202]: DEBUG nova.scheduler.client.report [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 844.357387] env[63202]: INFO nova.scheduler.client.report [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Deleted allocations for instance 06bd8147-5479-4c5e-8ba2-6d3a205ba05e [ 844.565285] env[63202]: DEBUG nova.network.neutron [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.656031] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquiring lock "7a72a1ab-3b3d-450d-a3a6-572b0035eb0e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.656256] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lock "7a72a1ab-3b3d-450d-a3a6-572b0035eb0e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.677213] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.828s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.677839] env[63202]: ERROR nova.compute.manager [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 115c9b60-ce54-4ffa-9bb5-eae100911988, please check neutron logs for more information. [ 844.677839] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] Traceback (most recent call last): [ 844.677839] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 844.677839] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] self.driver.spawn(context, instance, image_meta, [ 844.677839] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 844.677839] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 844.677839] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 844.677839] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] vm_ref = self.build_virtual_machine(instance, [ 844.677839] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 844.677839] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] vif_infos = vmwarevif.get_vif_info(self._session, [ 844.677839] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 844.678249] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] for vif in network_info: [ 844.678249] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 844.678249] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] return self._sync_wrapper(fn, *args, **kwargs) [ 844.678249] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 844.678249] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] self.wait() [ 844.678249] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 844.678249] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] self[:] = self._gt.wait() [ 844.678249] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 844.678249] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] return self._exit_event.wait() [ 844.678249] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 844.678249] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] result = hub.switch() [ 844.678249] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 844.678249] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] return self.greenlet.switch() [ 844.678738] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 844.678738] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] result = function(*args, **kwargs) [ 844.678738] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 844.678738] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] return func(*args, **kwargs) [ 844.678738] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 844.678738] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] raise e [ 844.678738] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 844.678738] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] nwinfo = self.network_api.allocate_for_instance( [ 844.678738] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 844.678738] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] created_port_ids = self._update_ports_for_instance( [ 844.678738] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 844.678738] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] with excutils.save_and_reraise_exception(): [ 844.678738] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 844.679264] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] self.force_reraise() [ 844.679264] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 844.679264] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] raise self.value [ 844.679264] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 844.679264] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] updated_port = self._update_port( [ 844.679264] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 844.679264] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] _ensure_no_port_binding_failure(port) [ 844.679264] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 844.679264] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] raise exception.PortBindingFailed(port_id=port['id']) [ 844.679264] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] nova.exception.PortBindingFailed: Binding failed for port 115c9b60-ce54-4ffa-9bb5-eae100911988, please check neutron logs for more information. [ 844.679264] env[63202]: ERROR nova.compute.manager [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] [ 844.679589] env[63202]: DEBUG nova.compute.utils [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] Binding failed for port 115c9b60-ce54-4ffa-9bb5-eae100911988, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 844.679694] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.580s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.682677] env[63202]: DEBUG nova.compute.manager [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] Build of instance 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7 was re-scheduled: Binding failed for port 115c9b60-ce54-4ffa-9bb5-eae100911988, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 844.683112] env[63202]: DEBUG nova.compute.manager [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 844.683330] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "refresh_cache-0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.683473] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquired lock "refresh_cache-0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.683628] env[63202]: DEBUG nova.network.neutron [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 844.867178] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89d62ac2-4799-47d8-8b97-bc624c3e0cc8 tempest-SecurityGroupsTestJSON-1001345364 tempest-SecurityGroupsTestJSON-1001345364-project-member] Lock "06bd8147-5479-4c5e-8ba2-6d3a205ba05e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 159.052s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.069991] env[63202]: INFO nova.compute.manager [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] [instance: 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a] Took 1.02 seconds to deallocate network for instance. [ 845.211503] env[63202]: DEBUG nova.network.neutron [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 845.359210] env[63202]: DEBUG nova.network.neutron [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.368926] env[63202]: DEBUG nova.compute.manager [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 845.437023] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf737a56-b302-420e-87a3-ae379121e002 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.444973] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bd7a3ce-a378-4de7-a1ca-40c918225fdc {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.474018] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a86d53e-6f31-47ac-98d1-f3c5d37ee151 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.481737] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b301ec3-7d35-4893-9958-c72eb4038ae1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.495398] env[63202]: DEBUG nova.compute.provider_tree [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 845.864182] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Releasing lock "refresh_cache-0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.864182] env[63202]: DEBUG nova.compute.manager [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 845.864182] env[63202]: DEBUG nova.compute.manager [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 845.864182] env[63202]: DEBUG nova.network.neutron [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 845.885453] env[63202]: DEBUG nova.network.neutron [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 845.892273] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.001222] env[63202]: DEBUG nova.scheduler.client.report [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 846.098713] env[63202]: INFO nova.scheduler.client.report [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] Deleted allocations for instance 56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a [ 846.388029] env[63202]: DEBUG nova.network.neutron [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.505843] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.826s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.506506] env[63202]: ERROR nova.compute.manager [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9f87d11f-db0c-4b6b-84a2-86c00886f0db, please check neutron logs for more information. [ 846.506506] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] Traceback (most recent call last): [ 846.506506] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 846.506506] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] self.driver.spawn(context, instance, image_meta, [ 846.506506] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 846.506506] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 846.506506] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 846.506506] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] vm_ref = self.build_virtual_machine(instance, [ 846.506506] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 846.506506] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] vif_infos = vmwarevif.get_vif_info(self._session, [ 846.506506] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 846.506802] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] for vif in network_info: [ 846.506802] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 846.506802] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] return self._sync_wrapper(fn, *args, **kwargs) [ 846.506802] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 846.506802] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] self.wait() [ 846.506802] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 846.506802] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] self[:] = self._gt.wait() [ 846.506802] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 846.506802] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] return self._exit_event.wait() [ 846.506802] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 846.506802] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] current.throw(*self._exc) [ 846.506802] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 846.506802] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] result = function(*args, **kwargs) [ 846.507207] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 846.507207] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] return func(*args, **kwargs) [ 846.507207] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 846.507207] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] raise e [ 846.507207] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 846.507207] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] nwinfo = self.network_api.allocate_for_instance( [ 846.507207] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 846.507207] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] created_port_ids = self._update_ports_for_instance( [ 846.507207] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 846.507207] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] with excutils.save_and_reraise_exception(): [ 846.507207] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 846.507207] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] self.force_reraise() [ 846.507207] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 846.507514] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] raise self.value [ 846.507514] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 846.507514] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] updated_port = self._update_port( [ 846.507514] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 846.507514] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] _ensure_no_port_binding_failure(port) [ 846.507514] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 846.507514] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] raise exception.PortBindingFailed(port_id=port['id']) [ 846.507514] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] nova.exception.PortBindingFailed: Binding failed for port 9f87d11f-db0c-4b6b-84a2-86c00886f0db, please check neutron logs for more information. [ 846.507514] env[63202]: ERROR nova.compute.manager [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] [ 846.507514] env[63202]: DEBUG nova.compute.utils [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] Binding failed for port 9f87d11f-db0c-4b6b-84a2-86c00886f0db, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 846.508584] env[63202]: DEBUG oslo_concurrency.lockutils [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.895s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.511734] env[63202]: DEBUG nova.compute.manager [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] Build of instance a47dbb3e-62be-426d-acd7-bedc0faf65c5 was re-scheduled: Binding failed for port 9f87d11f-db0c-4b6b-84a2-86c00886f0db, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 846.512216] env[63202]: DEBUG nova.compute.manager [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 846.512441] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] Acquiring lock "refresh_cache-a47dbb3e-62be-426d-acd7-bedc0faf65c5" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.512587] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] Acquired lock "refresh_cache-a47dbb3e-62be-426d-acd7-bedc0faf65c5" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.512741] env[63202]: DEBUG nova.network.neutron [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 846.605924] env[63202]: DEBUG oslo_concurrency.lockutils [None req-80ba9e4e-4977-45f2-9c67-ec3e513e394a tempest-ServerDiagnosticsNegativeTest-1635561772 tempest-ServerDiagnosticsNegativeTest-1635561772-project-member] Lock "56feb1b8-7dbb-4a5c-a8e9-5cb53059e39a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 160.440s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.890701] env[63202]: INFO nova.compute.manager [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7] Took 1.03 seconds to deallocate network for instance. [ 847.060709] env[63202]: DEBUG nova.network.neutron [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 847.110216] env[63202]: DEBUG nova.compute.manager [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 847.219147] env[63202]: DEBUG nova.network.neutron [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.331745] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfecfa65-04d0-4110-8ddf-190732a1d306 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.339702] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1361850d-cdf0-42ee-ad78-f5f46b9cdfdc {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.369537] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2535ad52-8935-4da4-8f60-5ca8e6fe4e37 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.377062] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63a13729-a259-47b2-9bbb-c84d7cd8af0e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.391961] env[63202]: DEBUG nova.compute.provider_tree [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 847.637714] env[63202]: DEBUG oslo_concurrency.lockutils [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.722618] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] Releasing lock "refresh_cache-a47dbb3e-62be-426d-acd7-bedc0faf65c5" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.722910] env[63202]: DEBUG nova.compute.manager [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 847.723109] env[63202]: DEBUG nova.compute.manager [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 847.723278] env[63202]: DEBUG nova.network.neutron [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 847.748592] env[63202]: DEBUG nova.network.neutron [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 847.895931] env[63202]: DEBUG nova.scheduler.client.report [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 847.921731] env[63202]: INFO nova.scheduler.client.report [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Deleted allocations for instance 0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7 [ 848.254642] env[63202]: DEBUG nova.network.neutron [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.400860] env[63202]: DEBUG oslo_concurrency.lockutils [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.892s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.401523] env[63202]: ERROR nova.compute.manager [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port cc6896d9-1f04-49c5-82f3-1d50d3ef5244, please check neutron logs for more information. [ 848.401523] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] Traceback (most recent call last): [ 848.401523] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 848.401523] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] self.driver.spawn(context, instance, image_meta, [ 848.401523] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 848.401523] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] self._vmops.spawn(context, instance, image_meta, injected_files, [ 848.401523] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 848.401523] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] vm_ref = self.build_virtual_machine(instance, [ 848.401523] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 848.401523] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] vif_infos = vmwarevif.get_vif_info(self._session, [ 848.401523] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 848.401839] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] for vif in network_info: [ 848.401839] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 848.401839] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] return self._sync_wrapper(fn, *args, **kwargs) [ 848.401839] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 848.401839] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] self.wait() [ 848.401839] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 848.401839] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] self[:] = self._gt.wait() [ 848.401839] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 848.401839] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] return self._exit_event.wait() [ 848.401839] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 848.401839] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] current.throw(*self._exc) [ 848.401839] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 848.401839] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] result = function(*args, **kwargs) [ 848.402163] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 848.402163] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] return func(*args, **kwargs) [ 848.402163] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 848.402163] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] raise e [ 848.402163] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 848.402163] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] nwinfo = self.network_api.allocate_for_instance( [ 848.402163] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 848.402163] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] created_port_ids = self._update_ports_for_instance( [ 848.402163] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 848.402163] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] with excutils.save_and_reraise_exception(): [ 848.402163] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 848.402163] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] self.force_reraise() [ 848.402163] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 848.402429] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] raise self.value [ 848.402429] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 848.402429] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] updated_port = self._update_port( [ 848.402429] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 848.402429] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] _ensure_no_port_binding_failure(port) [ 848.402429] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 848.402429] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] raise exception.PortBindingFailed(port_id=port['id']) [ 848.402429] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] nova.exception.PortBindingFailed: Binding failed for port cc6896d9-1f04-49c5-82f3-1d50d3ef5244, please check neutron logs for more information. [ 848.402429] env[63202]: ERROR nova.compute.manager [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] [ 848.402429] env[63202]: DEBUG nova.compute.utils [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] Binding failed for port cc6896d9-1f04-49c5-82f3-1d50d3ef5244, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 848.403435] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.248s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.406402] env[63202]: DEBUG nova.compute.manager [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] Build of instance 574ddcba-db96-4bd1-bb9e-3f8846c13430 was re-scheduled: Binding failed for port cc6896d9-1f04-49c5-82f3-1d50d3ef5244, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 848.406938] env[63202]: DEBUG nova.compute.manager [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 848.407202] env[63202]: DEBUG oslo_concurrency.lockutils [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Acquiring lock "refresh_cache-574ddcba-db96-4bd1-bb9e-3f8846c13430" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.407352] env[63202]: DEBUG oslo_concurrency.lockutils [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Acquired lock "refresh_cache-574ddcba-db96-4bd1-bb9e-3f8846c13430" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.407508] env[63202]: DEBUG nova.network.neutron [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 848.431802] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c1b6e994-cea6-43a7-9b24-eaead4e7aeed tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "0ea5fdd0-6f55-4e8f-b18b-261e7edd4cf7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 161.382s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.757788] env[63202]: INFO nova.compute.manager [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] [instance: a47dbb3e-62be-426d-acd7-bedc0faf65c5] Took 1.03 seconds to deallocate network for instance. [ 848.932775] env[63202]: DEBUG nova.network.neutron [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 848.937362] env[63202]: DEBUG nova.compute.manager [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] [instance: a800fa5d-46a1-4a96-8ea2-26e34b869aa6] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 849.095013] env[63202]: DEBUG nova.network.neutron [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 849.215572] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d95fa0d7-78f9-4887-b2f8-9696a8808d3c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.225437] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-426799c3-05a8-4f05-acbc-5c6b8ae52ca1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.259970] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93911e54-c24c-4b1b-bb7f-2faa488ebbdc {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.271894] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-761c00b7-5311-4d28-95b2-aa22af35425e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.286180] env[63202]: DEBUG nova.compute.provider_tree [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 849.465395] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.597878] env[63202]: DEBUG oslo_concurrency.lockutils [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Releasing lock "refresh_cache-574ddcba-db96-4bd1-bb9e-3f8846c13430" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.598132] env[63202]: DEBUG nova.compute.manager [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 849.598312] env[63202]: DEBUG nova.compute.manager [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 849.598488] env[63202]: DEBUG nova.network.neutron [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 849.630498] env[63202]: DEBUG nova.network.neutron [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 849.794315] env[63202]: DEBUG nova.scheduler.client.report [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 849.809398] env[63202]: INFO nova.scheduler.client.report [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] Deleted allocations for instance a47dbb3e-62be-426d-acd7-bedc0faf65c5 [ 850.133283] env[63202]: DEBUG nova.network.neutron [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.299568] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.896s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.300232] env[63202]: ERROR nova.compute.manager [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 67590287-a9ba-4e46-93ad-f26ea015e2e6, please check neutron logs for more information. [ 850.300232] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] Traceback (most recent call last): [ 850.300232] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 850.300232] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] self.driver.spawn(context, instance, image_meta, [ 850.300232] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 850.300232] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 850.300232] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 850.300232] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] vm_ref = self.build_virtual_machine(instance, [ 850.300232] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 850.300232] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] vif_infos = vmwarevif.get_vif_info(self._session, [ 850.300232] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 850.300556] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] for vif in network_info: [ 850.300556] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 850.300556] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] return self._sync_wrapper(fn, *args, **kwargs) [ 850.300556] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 850.300556] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] self.wait() [ 850.300556] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 850.300556] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] self[:] = self._gt.wait() [ 850.300556] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 850.300556] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] return self._exit_event.wait() [ 850.300556] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 850.300556] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] current.throw(*self._exc) [ 850.300556] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 850.300556] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] result = function(*args, **kwargs) [ 850.300842] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 850.300842] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] return func(*args, **kwargs) [ 850.300842] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 850.300842] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] raise e [ 850.300842] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 850.300842] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] nwinfo = self.network_api.allocate_for_instance( [ 850.300842] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 850.300842] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] created_port_ids = self._update_ports_for_instance( [ 850.300842] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 850.300842] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] with excutils.save_and_reraise_exception(): [ 850.300842] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 850.300842] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] self.force_reraise() [ 850.300842] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 850.301134] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] raise self.value [ 850.301134] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 850.301134] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] updated_port = self._update_port( [ 850.301134] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 850.301134] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] _ensure_no_port_binding_failure(port) [ 850.301134] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 850.301134] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] raise exception.PortBindingFailed(port_id=port['id']) [ 850.301134] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] nova.exception.PortBindingFailed: Binding failed for port 67590287-a9ba-4e46-93ad-f26ea015e2e6, please check neutron logs for more information. [ 850.301134] env[63202]: ERROR nova.compute.manager [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] [ 850.301134] env[63202]: DEBUG nova.compute.utils [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] Binding failed for port 67590287-a9ba-4e46-93ad-f26ea015e2e6, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 850.302903] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.910s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.304133] env[63202]: INFO nova.compute.claims [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 850.310277] env[63202]: DEBUG nova.compute.manager [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] Build of instance 57f446ee-edf3-40e7-96bf-8fdce62bb19f was re-scheduled: Binding failed for port 67590287-a9ba-4e46-93ad-f26ea015e2e6, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 850.310812] env[63202]: DEBUG nova.compute.manager [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 850.311136] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] Acquiring lock "refresh_cache-57f446ee-edf3-40e7-96bf-8fdce62bb19f" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.311345] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] Acquired lock "refresh_cache-57f446ee-edf3-40e7-96bf-8fdce62bb19f" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.311588] env[63202]: DEBUG nova.network.neutron [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 850.320727] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c63d80ed-0e32-4c6d-8d18-b1b4a5b00d91 tempest-ServerActionsTestOtherA-289050786 tempest-ServerActionsTestOtherA-289050786-project-member] Lock "a47dbb3e-62be-426d-acd7-bedc0faf65c5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 155.338s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.423280] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "e775e5e1-521a-4fc7-80e6-bcb6a70516c5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.423280] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "e775e5e1-521a-4fc7-80e6-bcb6a70516c5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.639129] env[63202]: INFO nova.compute.manager [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] [instance: 574ddcba-db96-4bd1-bb9e-3f8846c13430] Took 1.04 seconds to deallocate network for instance. [ 850.823444] env[63202]: DEBUG nova.compute.manager [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 850.836106] env[63202]: DEBUG nova.network.neutron [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 851.089331] env[63202]: DEBUG oslo_concurrency.lockutils [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Acquiring lock "753d190b-f4a4-4438-bc98-94564e3ec73d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.089562] env[63202]: DEBUG oslo_concurrency.lockutils [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Lock "753d190b-f4a4-4438-bc98-94564e3ec73d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.126352] env[63202]: DEBUG nova.network.neutron [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.350185] env[63202]: DEBUG oslo_concurrency.lockutils [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.608427] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ef9c5d3-b30b-437a-a567-b337b6f95394 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.619892] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eadb158-ccec-4677-8656-dc6933b2f889 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.650130] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] Releasing lock "refresh_cache-57f446ee-edf3-40e7-96bf-8fdce62bb19f" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.650376] env[63202]: DEBUG nova.compute.manager [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 851.650550] env[63202]: DEBUG nova.compute.manager [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 851.650705] env[63202]: DEBUG nova.network.neutron [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 851.655920] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d83e2efb-2dcb-4994-be97-8809123c1911 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.665108] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01d0966a-d4bd-414d-b099-526eabf1596a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.679910] env[63202]: DEBUG nova.compute.provider_tree [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 851.683995] env[63202]: INFO nova.scheduler.client.report [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Deleted allocations for instance 574ddcba-db96-4bd1-bb9e-3f8846c13430 [ 851.690182] env[63202]: DEBUG nova.network.neutron [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 852.192899] env[63202]: DEBUG nova.scheduler.client.report [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 852.198503] env[63202]: DEBUG nova.network.neutron [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.198503] env[63202]: DEBUG oslo_concurrency.lockutils [None req-517b96b0-133b-4413-8b5f-8ecfbfcd8afe tempest-AttachInterfacesTestJSON-8017105 tempest-AttachInterfacesTestJSON-8017105-project-member] Lock "574ddcba-db96-4bd1-bb9e-3f8846c13430" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 118.709s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.364312] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Acquiring lock "af2bedc2-28ee-4679-ae38-1cceb2af05d6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.364580] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Lock "af2bedc2-28ee-4679-ae38-1cceb2af05d6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.387708] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Acquiring lock "b1bccea9-2d79-431a-8be0-0a5ab293542a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.387997] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Lock "b1bccea9-2d79-431a-8be0-0a5ab293542a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.416064] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Acquiring lock "f3652744-e072-4700-80d4-b9eca414c5cb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.416313] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Lock "f3652744-e072-4700-80d4-b9eca414c5cb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.702254] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.397s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.702254] env[63202]: DEBUG nova.compute.manager [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 852.703596] env[63202]: INFO nova.compute.manager [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] [instance: 57f446ee-edf3-40e7-96bf-8fdce62bb19f] Took 1.05 seconds to deallocate network for instance. [ 852.706393] env[63202]: DEBUG nova.compute.manager [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 852.710114] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 13.470s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.218026] env[63202]: DEBUG nova.compute.utils [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 853.221549] env[63202]: DEBUG nova.compute.manager [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 853.221781] env[63202]: DEBUG nova.network.neutron [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 853.250281] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.318390] env[63202]: DEBUG nova.policy [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cae2051248b04d029f38a42784e72112', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9fe128a6db444bda82fa3a75137ad99a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 853.667838] env[63202]: DEBUG nova.network.neutron [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] Successfully created port: e9f72a2a-5ce9-4802-bd15-9c22c86662b9 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 853.727817] env[63202]: DEBUG nova.compute.manager [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 853.774037] env[63202]: INFO nova.scheduler.client.report [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] Deleted allocations for instance 57f446ee-edf3-40e7-96bf-8fdce62bb19f [ 854.265890] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 57f446ee-edf3-40e7-96bf-8fdce62bb19f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 854.266070] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance ac48e21a-72ec-4c9d-8262-5d6cb36ec531 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 854.286131] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1cfaeb60-fb6a-4466-8d83-d5b1f942587c tempest-ServerAddressesTestJSON-1475001474 tempest-ServerAddressesTestJSON-1475001474-project-member] Lock "57f446ee-edf3-40e7-96bf-8fdce62bb19f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 109.909s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.690722] env[63202]: DEBUG nova.compute.manager [req-67dd8edc-e41f-4c87-ad56-05fad93ca5f0 req-45ddeb7e-f5f0-47f7-acbd-a228ed667dcf service nova] [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] Received event network-changed-e9f72a2a-5ce9-4802-bd15-9c22c86662b9 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 854.691071] env[63202]: DEBUG nova.compute.manager [req-67dd8edc-e41f-4c87-ad56-05fad93ca5f0 req-45ddeb7e-f5f0-47f7-acbd-a228ed667dcf service nova] [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] Refreshing instance network info cache due to event network-changed-e9f72a2a-5ce9-4802-bd15-9c22c86662b9. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 854.691154] env[63202]: DEBUG oslo_concurrency.lockutils [req-67dd8edc-e41f-4c87-ad56-05fad93ca5f0 req-45ddeb7e-f5f0-47f7-acbd-a228ed667dcf service nova] Acquiring lock "refresh_cache-ac48e21a-72ec-4c9d-8262-5d6cb36ec531" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.691314] env[63202]: DEBUG oslo_concurrency.lockutils [req-67dd8edc-e41f-4c87-ad56-05fad93ca5f0 req-45ddeb7e-f5f0-47f7-acbd-a228ed667dcf service nova] Acquired lock "refresh_cache-ac48e21a-72ec-4c9d-8262-5d6cb36ec531" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.691492] env[63202]: DEBUG nova.network.neutron [req-67dd8edc-e41f-4c87-ad56-05fad93ca5f0 req-45ddeb7e-f5f0-47f7-acbd-a228ed667dcf service nova] [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] Refreshing network info cache for port e9f72a2a-5ce9-4802-bd15-9c22c86662b9 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 854.738944] env[63202]: DEBUG nova.compute.manager [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 854.772101] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 69c66a35-3789-4c2f-b7cf-5f4136d53b94 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 854.776199] env[63202]: DEBUG nova.virt.hardware [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 854.776424] env[63202]: DEBUG nova.virt.hardware [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 854.776577] env[63202]: DEBUG nova.virt.hardware [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 854.776756] env[63202]: DEBUG nova.virt.hardware [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 854.776893] env[63202]: DEBUG nova.virt.hardware [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 854.777044] env[63202]: DEBUG nova.virt.hardware [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 854.777249] env[63202]: DEBUG nova.virt.hardware [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 854.777405] env[63202]: DEBUG nova.virt.hardware [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 854.777596] env[63202]: DEBUG nova.virt.hardware [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 854.777760] env[63202]: DEBUG nova.virt.hardware [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 854.777929] env[63202]: DEBUG nova.virt.hardware [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 854.783705] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e1f6f79-ccc8-44d7-aadd-2fb9725fd788 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.788962] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6d77c06-2975-43e5-9a0d-7c41921ec801 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.794363] env[63202]: DEBUG nova.compute.manager [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 854.963030] env[63202]: ERROR nova.compute.manager [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e9f72a2a-5ce9-4802-bd15-9c22c86662b9, please check neutron logs for more information. [ 854.963030] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 854.963030] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 854.963030] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 854.963030] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 854.963030] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 854.963030] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 854.963030] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 854.963030] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 854.963030] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 854.963030] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 854.963030] env[63202]: ERROR nova.compute.manager raise self.value [ 854.963030] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 854.963030] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 854.963030] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 854.963030] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 854.963437] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 854.963437] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 854.963437] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e9f72a2a-5ce9-4802-bd15-9c22c86662b9, please check neutron logs for more information. [ 854.963437] env[63202]: ERROR nova.compute.manager [ 854.963437] env[63202]: Traceback (most recent call last): [ 854.963437] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 854.963437] env[63202]: listener.cb(fileno) [ 854.963437] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 854.963437] env[63202]: result = function(*args, **kwargs) [ 854.963437] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 854.963437] env[63202]: return func(*args, **kwargs) [ 854.963437] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 854.963437] env[63202]: raise e [ 854.963437] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 854.963437] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 854.963437] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 854.963437] env[63202]: created_port_ids = self._update_ports_for_instance( [ 854.963437] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 854.963437] env[63202]: with excutils.save_and_reraise_exception(): [ 854.963437] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 854.963437] env[63202]: self.force_reraise() [ 854.963437] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 854.963437] env[63202]: raise self.value [ 854.963437] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 854.963437] env[63202]: updated_port = self._update_port( [ 854.963437] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 854.963437] env[63202]: _ensure_no_port_binding_failure(port) [ 854.963437] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 854.963437] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 854.964230] env[63202]: nova.exception.PortBindingFailed: Binding failed for port e9f72a2a-5ce9-4802-bd15-9c22c86662b9, please check neutron logs for more information. [ 854.964230] env[63202]: Removing descriptor: 16 [ 854.964230] env[63202]: ERROR nova.compute.manager [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e9f72a2a-5ce9-4802-bd15-9c22c86662b9, please check neutron logs for more information. [ 854.964230] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] Traceback (most recent call last): [ 854.964230] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 854.964230] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] yield resources [ 854.964230] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 854.964230] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] self.driver.spawn(context, instance, image_meta, [ 854.964230] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 854.964230] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] self._vmops.spawn(context, instance, image_meta, injected_files, [ 854.964230] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 854.964230] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] vm_ref = self.build_virtual_machine(instance, [ 854.964537] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 854.964537] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] vif_infos = vmwarevif.get_vif_info(self._session, [ 854.964537] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 854.964537] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] for vif in network_info: [ 854.964537] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 854.964537] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] return self._sync_wrapper(fn, *args, **kwargs) [ 854.964537] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 854.964537] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] self.wait() [ 854.964537] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 854.964537] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] self[:] = self._gt.wait() [ 854.964537] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 854.964537] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] return self._exit_event.wait() [ 854.964537] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 854.964813] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] result = hub.switch() [ 854.964813] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 854.964813] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] return self.greenlet.switch() [ 854.964813] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 854.964813] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] result = function(*args, **kwargs) [ 854.964813] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 854.964813] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] return func(*args, **kwargs) [ 854.964813] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 854.964813] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] raise e [ 854.964813] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 854.964813] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] nwinfo = self.network_api.allocate_for_instance( [ 854.964813] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 854.964813] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] created_port_ids = self._update_ports_for_instance( [ 854.965096] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 854.965096] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] with excutils.save_and_reraise_exception(): [ 854.965096] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 854.965096] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] self.force_reraise() [ 854.965096] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 854.965096] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] raise self.value [ 854.965096] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 854.965096] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] updated_port = self._update_port( [ 854.965096] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 854.965096] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] _ensure_no_port_binding_failure(port) [ 854.965096] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 854.965096] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] raise exception.PortBindingFailed(port_id=port['id']) [ 854.965358] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] nova.exception.PortBindingFailed: Binding failed for port e9f72a2a-5ce9-4802-bd15-9c22c86662b9, please check neutron logs for more information. [ 854.965358] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] [ 854.965358] env[63202]: INFO nova.compute.manager [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] Terminating instance [ 854.968205] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] Acquiring lock "refresh_cache-ac48e21a-72ec-4c9d-8262-5d6cb36ec531" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.222736] env[63202]: DEBUG nova.network.neutron [req-67dd8edc-e41f-4c87-ad56-05fad93ca5f0 req-45ddeb7e-f5f0-47f7-acbd-a228ed667dcf service nova] [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 855.274940] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 5d9f1803-e3f8-42a5-b6b7-2642463a079e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 855.326471] env[63202]: DEBUG oslo_concurrency.lockutils [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.335561] env[63202]: DEBUG nova.network.neutron [req-67dd8edc-e41f-4c87-ad56-05fad93ca5f0 req-45ddeb7e-f5f0-47f7-acbd-a228ed667dcf service nova] [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.782035] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance d674db7a-8cf0-46f8-b676-fced0767c105 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 855.838870] env[63202]: DEBUG oslo_concurrency.lockutils [req-67dd8edc-e41f-4c87-ad56-05fad93ca5f0 req-45ddeb7e-f5f0-47f7-acbd-a228ed667dcf service nova] Releasing lock "refresh_cache-ac48e21a-72ec-4c9d-8262-5d6cb36ec531" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.839164] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] Acquired lock "refresh_cache-ac48e21a-72ec-4c9d-8262-5d6cb36ec531" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.839348] env[63202]: DEBUG nova.network.neutron [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 856.285889] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance cedc3a06-2123-4c5e-a6c3-599a3efc3c65 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 856.366838] env[63202]: DEBUG nova.network.neutron [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 856.499010] env[63202]: DEBUG nova.network.neutron [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.721674] env[63202]: DEBUG nova.compute.manager [req-e4cedb23-befd-48dc-9c69-c5292c91eec7 req-f06c7b77-88f2-46c1-a43c-7cdd019899bd service nova] [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] Received event network-vif-deleted-e9f72a2a-5ce9-4802-bd15-9c22c86662b9 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 856.794787] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 857.012375] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] Releasing lock "refresh_cache-ac48e21a-72ec-4c9d-8262-5d6cb36ec531" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.012801] env[63202]: DEBUG nova.compute.manager [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 857.012994] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 857.013338] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-900d5f67-46a8-4e10-8475-1b28514b5a8e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.027858] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8ebf916-59cf-493c-ab09-4c9604d47cd1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.057520] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ac48e21a-72ec-4c9d-8262-5d6cb36ec531 could not be found. [ 857.057799] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 857.057983] env[63202]: INFO nova.compute.manager [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] Took 0.04 seconds to destroy the instance on the hypervisor. [ 857.058300] env[63202]: DEBUG oslo.service.loopingcall [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 857.058439] env[63202]: DEBUG nova.compute.manager [-] [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 857.058527] env[63202]: DEBUG nova.network.neutron [-] [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 857.076271] env[63202]: DEBUG nova.network.neutron [-] [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 857.301936] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance a800fa5d-46a1-4a96-8ea2-26e34b869aa6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 857.578767] env[63202]: DEBUG nova.network.neutron [-] [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.807376] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 857.945779] env[63202]: DEBUG oslo_concurrency.lockutils [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Acquiring lock "fc7d3924-a624-4ea4-890c-7628595fb733" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.945886] env[63202]: DEBUG oslo_concurrency.lockutils [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Lock "fc7d3924-a624-4ea4-890c-7628595fb733" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.083912] env[63202]: INFO nova.compute.manager [-] [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] Took 1.02 seconds to deallocate network for instance. [ 858.086430] env[63202]: DEBUG nova.compute.claims [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 858.086624] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.315257] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 50854a4e-8e4e-431b-8220-01bf8906b1c0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 858.524471] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Acquiring lock "d0d6e380-9337-4f69-8434-6a33ac8a33d6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.524706] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Lock "d0d6e380-9337-4f69-8434-6a33ac8a33d6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.818401] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 2243bb69-0dc5-49cd-b94e-73e703cbadc0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 859.321098] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 36291628-af7d-43cf-8149-09f57df47890 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 859.824705] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 3d1518d3-bed2-4b7b-af1d-d7da49e92874 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 860.327874] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 61a6528e-4e4e-49b7-be7b-dd30bd8023d8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 860.831350] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 861.334011] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance e775e5e1-521a-4fc7-80e6-bcb6a70516c5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 861.836939] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 753d190b-f4a4-4438-bc98-94564e3ec73d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 862.340019] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance af2bedc2-28ee-4679-ae38-1cceb2af05d6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 862.843353] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance b1bccea9-2d79-431a-8be0-0a5ab293542a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 863.346623] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance f3652744-e072-4700-80d4-b9eca414c5cb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 863.346860] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=63202) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 863.347014] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=63202) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 863.591838] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e96246a-9036-41f9-af1a-a97b62cdad45 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.600792] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04fd51f2-3579-4d37-9b9e-5d6229eb741c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.629391] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e319e5a-9a81-4c3d-bbf5-df65548869b4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.636525] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d35a9d1-b0c2-4f0b-a68d-8def761e0228 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.649999] env[63202]: DEBUG nova.compute.provider_tree [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 864.153662] env[63202]: DEBUG nova.scheduler.client.report [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 864.660051] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63202) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 864.660308] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 11.950s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.660638] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.491s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.662789] env[63202]: INFO nova.compute.claims [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 865.630293] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 865.630636] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Starting heal instance info cache {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 865.630636] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Rebuilding the list of instances to heal {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 865.887750] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6432649b-b4ca-4e49-ae58-1b6fad097e1d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.895251] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a71067f-9b86-4b5a-ac8b-c29b064661fe {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.923956] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51412c4e-eeed-45ce-b22f-01712e35c0d4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.930497] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1540a42-7bd4-40ec-ba92-5c1e28df3d3f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.943609] env[63202]: DEBUG nova.compute.provider_tree [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 866.135651] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] Skipping network cache update for instance because it is Building. {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 866.135651] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] Skipping network cache update for instance because it is Building. {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 866.135765] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Didn't find any instances for network info cache update. {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 866.135944] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 866.447110] env[63202]: DEBUG nova.scheduler.client.report [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 866.952032] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.291s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.952652] env[63202]: DEBUG nova.compute.manager [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 866.955469] env[63202]: DEBUG oslo_concurrency.lockutils [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.792s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.957015] env[63202]: INFO nova.compute.claims [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 867.462303] env[63202]: DEBUG nova.compute.utils [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 867.465697] env[63202]: DEBUG nova.compute.manager [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 867.465869] env[63202]: DEBUG nova.network.neutron [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 867.526776] env[63202]: DEBUG nova.policy [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '13a4740a28d64585978b97b5f29c14a3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a16afda75a3f4126820883cd01e4e459', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 867.836985] env[63202]: DEBUG nova.network.neutron [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] Successfully created port: 6b7258a2-07d2-4cb5-b458-04ed148d1895 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 867.967097] env[63202]: DEBUG nova.compute.manager [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 868.258363] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4188f08-32ab-41b3-9509-59d0ccc26243 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.265936] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-192178cc-7a2c-4c86-bfc0-7043be97f57e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.296610] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c997ea6-8d37-41ba-a369-7e539a958969 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.303863] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f35faff9-d605-492f-8eae-a077960aa18b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.316888] env[63202]: DEBUG nova.compute.provider_tree [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 868.680515] env[63202]: DEBUG nova.compute.manager [req-55f7d316-9f6d-42d8-b09c-792c46de5a16 req-b1e65c31-6d25-40f0-b6ee-d753bab049a0 service nova] [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] Received event network-changed-6b7258a2-07d2-4cb5-b458-04ed148d1895 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 868.680742] env[63202]: DEBUG nova.compute.manager [req-55f7d316-9f6d-42d8-b09c-792c46de5a16 req-b1e65c31-6d25-40f0-b6ee-d753bab049a0 service nova] [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] Refreshing instance network info cache due to event network-changed-6b7258a2-07d2-4cb5-b458-04ed148d1895. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 868.680964] env[63202]: DEBUG oslo_concurrency.lockutils [req-55f7d316-9f6d-42d8-b09c-792c46de5a16 req-b1e65c31-6d25-40f0-b6ee-d753bab049a0 service nova] Acquiring lock "refresh_cache-69c66a35-3789-4c2f-b7cf-5f4136d53b94" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.681127] env[63202]: DEBUG oslo_concurrency.lockutils [req-55f7d316-9f6d-42d8-b09c-792c46de5a16 req-b1e65c31-6d25-40f0-b6ee-d753bab049a0 service nova] Acquired lock "refresh_cache-69c66a35-3789-4c2f-b7cf-5f4136d53b94" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.681339] env[63202]: DEBUG nova.network.neutron [req-55f7d316-9f6d-42d8-b09c-792c46de5a16 req-b1e65c31-6d25-40f0-b6ee-d753bab049a0 service nova] [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] Refreshing network info cache for port 6b7258a2-07d2-4cb5-b458-04ed148d1895 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 868.821117] env[63202]: DEBUG nova.scheduler.client.report [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 868.891176] env[63202]: ERROR nova.compute.manager [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6b7258a2-07d2-4cb5-b458-04ed148d1895, please check neutron logs for more information. [ 868.891176] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 868.891176] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 868.891176] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 868.891176] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 868.891176] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 868.891176] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 868.891176] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 868.891176] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 868.891176] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 868.891176] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 868.891176] env[63202]: ERROR nova.compute.manager raise self.value [ 868.891176] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 868.891176] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 868.891176] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 868.891176] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 868.892036] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 868.892036] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 868.892036] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6b7258a2-07d2-4cb5-b458-04ed148d1895, please check neutron logs for more information. [ 868.892036] env[63202]: ERROR nova.compute.manager [ 868.892036] env[63202]: Traceback (most recent call last): [ 868.892036] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 868.892036] env[63202]: listener.cb(fileno) [ 868.892036] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 868.892036] env[63202]: result = function(*args, **kwargs) [ 868.892036] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 868.892036] env[63202]: return func(*args, **kwargs) [ 868.892036] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 868.892036] env[63202]: raise e [ 868.892036] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 868.892036] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 868.892036] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 868.892036] env[63202]: created_port_ids = self._update_ports_for_instance( [ 868.892036] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 868.892036] env[63202]: with excutils.save_and_reraise_exception(): [ 868.892036] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 868.892036] env[63202]: self.force_reraise() [ 868.892036] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 868.892036] env[63202]: raise self.value [ 868.892036] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 868.892036] env[63202]: updated_port = self._update_port( [ 868.892036] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 868.892036] env[63202]: _ensure_no_port_binding_failure(port) [ 868.892036] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 868.892036] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 868.893169] env[63202]: nova.exception.PortBindingFailed: Binding failed for port 6b7258a2-07d2-4cb5-b458-04ed148d1895, please check neutron logs for more information. [ 868.893169] env[63202]: Removing descriptor: 16 [ 868.979132] env[63202]: DEBUG nova.compute.manager [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 869.003338] env[63202]: DEBUG nova.virt.hardware [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 869.003595] env[63202]: DEBUG nova.virt.hardware [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 869.003750] env[63202]: DEBUG nova.virt.hardware [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 869.003926] env[63202]: DEBUG nova.virt.hardware [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 869.004082] env[63202]: DEBUG nova.virt.hardware [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 869.004231] env[63202]: DEBUG nova.virt.hardware [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 869.004437] env[63202]: DEBUG nova.virt.hardware [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 869.004591] env[63202]: DEBUG nova.virt.hardware [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 869.004754] env[63202]: DEBUG nova.virt.hardware [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 869.004913] env[63202]: DEBUG nova.virt.hardware [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 869.005097] env[63202]: DEBUG nova.virt.hardware [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 869.005941] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6fdfbb2-8e59-48aa-b745-14322528f285 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.013782] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6787081b-2e67-4b3e-9f96-4072679ca97b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.027533] env[63202]: ERROR nova.compute.manager [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6b7258a2-07d2-4cb5-b458-04ed148d1895, please check neutron logs for more information. [ 869.027533] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] Traceback (most recent call last): [ 869.027533] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 869.027533] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] yield resources [ 869.027533] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 869.027533] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] self.driver.spawn(context, instance, image_meta, [ 869.027533] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 869.027533] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] self._vmops.spawn(context, instance, image_meta, injected_files, [ 869.027533] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 869.027533] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] vm_ref = self.build_virtual_machine(instance, [ 869.027533] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 869.027848] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] vif_infos = vmwarevif.get_vif_info(self._session, [ 869.027848] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 869.027848] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] for vif in network_info: [ 869.027848] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 869.027848] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] return self._sync_wrapper(fn, *args, **kwargs) [ 869.027848] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 869.027848] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] self.wait() [ 869.027848] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 869.027848] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] self[:] = self._gt.wait() [ 869.027848] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 869.027848] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] return self._exit_event.wait() [ 869.027848] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 869.027848] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] current.throw(*self._exc) [ 869.028216] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 869.028216] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] result = function(*args, **kwargs) [ 869.028216] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 869.028216] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] return func(*args, **kwargs) [ 869.028216] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 869.028216] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] raise e [ 869.028216] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 869.028216] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] nwinfo = self.network_api.allocate_for_instance( [ 869.028216] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 869.028216] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] created_port_ids = self._update_ports_for_instance( [ 869.028216] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 869.028216] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] with excutils.save_and_reraise_exception(): [ 869.028216] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 869.028517] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] self.force_reraise() [ 869.028517] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 869.028517] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] raise self.value [ 869.028517] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 869.028517] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] updated_port = self._update_port( [ 869.028517] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 869.028517] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] _ensure_no_port_binding_failure(port) [ 869.028517] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 869.028517] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] raise exception.PortBindingFailed(port_id=port['id']) [ 869.028517] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] nova.exception.PortBindingFailed: Binding failed for port 6b7258a2-07d2-4cb5-b458-04ed148d1895, please check neutron logs for more information. [ 869.028517] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] [ 869.028517] env[63202]: INFO nova.compute.manager [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] Terminating instance [ 869.030233] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] Acquiring lock "refresh_cache-69c66a35-3789-4c2f-b7cf-5f4136d53b94" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.199208] env[63202]: DEBUG nova.network.neutron [req-55f7d316-9f6d-42d8-b09c-792c46de5a16 req-b1e65c31-6d25-40f0-b6ee-d753bab049a0 service nova] [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 869.299440] env[63202]: DEBUG nova.network.neutron [req-55f7d316-9f6d-42d8-b09c-792c46de5a16 req-b1e65c31-6d25-40f0-b6ee-d753bab049a0 service nova] [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.325959] env[63202]: DEBUG oslo_concurrency.lockutils [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.370s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.326469] env[63202]: DEBUG nova.compute.manager [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 869.328862] env[63202]: DEBUG oslo_concurrency.lockutils [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.385s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.330577] env[63202]: INFO nova.compute.claims [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: d674db7a-8cf0-46f8-b676-fced0767c105] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 869.802303] env[63202]: DEBUG oslo_concurrency.lockutils [req-55f7d316-9f6d-42d8-b09c-792c46de5a16 req-b1e65c31-6d25-40f0-b6ee-d753bab049a0 service nova] Releasing lock "refresh_cache-69c66a35-3789-4c2f-b7cf-5f4136d53b94" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.802722] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] Acquired lock "refresh_cache-69c66a35-3789-4c2f-b7cf-5f4136d53b94" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.802901] env[63202]: DEBUG nova.network.neutron [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 869.834290] env[63202]: DEBUG nova.compute.utils [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 869.837331] env[63202]: DEBUG nova.compute.manager [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 869.837457] env[63202]: DEBUG nova.network.neutron [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 869.886755] env[63202]: DEBUG nova.policy [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '569158eee09b4bd597a32db0c2f02948', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '43717c26a67e4ea7bc9439ef2475b794', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 870.166303] env[63202]: DEBUG nova.network.neutron [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] Successfully created port: 55d081d1-5a3b-4a11-9c16-b73b57c34ea3 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 870.322761] env[63202]: DEBUG nova.network.neutron [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 870.338154] env[63202]: DEBUG nova.compute.manager [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 870.413479] env[63202]: DEBUG nova.network.neutron [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.694279] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d6b5bf3-65aa-4b00-92ba-dc84ebbffabc {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.702107] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-157b5cce-5cdc-4f12-88e8-4501dbbfa027 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.707334] env[63202]: DEBUG nova.compute.manager [req-3e23a755-8db4-48d9-9ce3-6a51bca65f6b req-0e3e5116-7490-45ab-b53d-f779a27d1d96 service nova] [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] Received event network-vif-deleted-6b7258a2-07d2-4cb5-b458-04ed148d1895 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 870.735130] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a781495-6472-4eec-967c-d5eb55e97460 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.741785] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d91f4f4-e904-47d5-93eb-b974940245c3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.755107] env[63202]: DEBUG nova.compute.provider_tree [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 870.916571] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] Releasing lock "refresh_cache-69c66a35-3789-4c2f-b7cf-5f4136d53b94" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.917213] env[63202]: DEBUG nova.compute.manager [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 870.918683] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 870.918683] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0ca72625-1657-493c-bd43-82c71e1f805f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.930221] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77660fc9-ed6a-46db-bca3-ac4775133a32 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.956922] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 69c66a35-3789-4c2f-b7cf-5f4136d53b94 could not be found. [ 870.956922] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 870.956922] env[63202]: INFO nova.compute.manager [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] Took 0.04 seconds to destroy the instance on the hypervisor. [ 870.957127] env[63202]: DEBUG oslo.service.loopingcall [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 870.957328] env[63202]: DEBUG nova.compute.manager [-] [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 870.957414] env[63202]: DEBUG nova.network.neutron [-] [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 870.975288] env[63202]: DEBUG nova.network.neutron [-] [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 871.258185] env[63202]: DEBUG nova.scheduler.client.report [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 871.349723] env[63202]: DEBUG nova.compute.manager [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 871.383183] env[63202]: DEBUG nova.virt.hardware [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 871.383442] env[63202]: DEBUG nova.virt.hardware [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 871.383601] env[63202]: DEBUG nova.virt.hardware [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 871.383774] env[63202]: DEBUG nova.virt.hardware [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 871.383914] env[63202]: DEBUG nova.virt.hardware [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 871.384288] env[63202]: DEBUG nova.virt.hardware [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 871.384533] env[63202]: DEBUG nova.virt.hardware [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 871.384695] env[63202]: DEBUG nova.virt.hardware [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 871.384998] env[63202]: DEBUG nova.virt.hardware [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 871.385193] env[63202]: DEBUG nova.virt.hardware [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 871.385367] env[63202]: DEBUG nova.virt.hardware [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 871.386226] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18c3784a-6463-472f-8da6-6d252c3ad74c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.394326] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ff92714-a2d8-4b6d-bb11-64e4a28539c7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.417040] env[63202]: ERROR nova.compute.manager [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 55d081d1-5a3b-4a11-9c16-b73b57c34ea3, please check neutron logs for more information. [ 871.417040] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 871.417040] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 871.417040] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 871.417040] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 871.417040] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 871.417040] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 871.417040] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 871.417040] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 871.417040] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 871.417040] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 871.417040] env[63202]: ERROR nova.compute.manager raise self.value [ 871.417040] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 871.417040] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 871.417040] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 871.417040] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 871.417827] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 871.417827] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 871.417827] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 55d081d1-5a3b-4a11-9c16-b73b57c34ea3, please check neutron logs for more information. [ 871.417827] env[63202]: ERROR nova.compute.manager [ 871.417827] env[63202]: Traceback (most recent call last): [ 871.417827] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 871.417827] env[63202]: listener.cb(fileno) [ 871.417827] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 871.417827] env[63202]: result = function(*args, **kwargs) [ 871.417827] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 871.417827] env[63202]: return func(*args, **kwargs) [ 871.417827] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 871.417827] env[63202]: raise e [ 871.417827] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 871.417827] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 871.417827] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 871.417827] env[63202]: created_port_ids = self._update_ports_for_instance( [ 871.417827] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 871.417827] env[63202]: with excutils.save_and_reraise_exception(): [ 871.417827] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 871.417827] env[63202]: self.force_reraise() [ 871.417827] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 871.417827] env[63202]: raise self.value [ 871.417827] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 871.417827] env[63202]: updated_port = self._update_port( [ 871.417827] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 871.417827] env[63202]: _ensure_no_port_binding_failure(port) [ 871.417827] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 871.417827] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 871.418595] env[63202]: nova.exception.PortBindingFailed: Binding failed for port 55d081d1-5a3b-4a11-9c16-b73b57c34ea3, please check neutron logs for more information. [ 871.418595] env[63202]: Removing descriptor: 16 [ 871.418595] env[63202]: ERROR nova.compute.manager [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 55d081d1-5a3b-4a11-9c16-b73b57c34ea3, please check neutron logs for more information. [ 871.418595] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] Traceback (most recent call last): [ 871.418595] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 871.418595] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] yield resources [ 871.418595] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 871.418595] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] self.driver.spawn(context, instance, image_meta, [ 871.418595] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 871.418595] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 871.418595] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 871.418595] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] vm_ref = self.build_virtual_machine(instance, [ 871.419021] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 871.419021] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] vif_infos = vmwarevif.get_vif_info(self._session, [ 871.419021] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 871.419021] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] for vif in network_info: [ 871.419021] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 871.419021] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] return self._sync_wrapper(fn, *args, **kwargs) [ 871.419021] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 871.419021] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] self.wait() [ 871.419021] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 871.419021] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] self[:] = self._gt.wait() [ 871.419021] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 871.419021] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] return self._exit_event.wait() [ 871.419021] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 871.419373] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] result = hub.switch() [ 871.419373] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 871.419373] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] return self.greenlet.switch() [ 871.419373] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 871.419373] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] result = function(*args, **kwargs) [ 871.419373] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 871.419373] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] return func(*args, **kwargs) [ 871.419373] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 871.419373] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] raise e [ 871.419373] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 871.419373] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] nwinfo = self.network_api.allocate_for_instance( [ 871.419373] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 871.419373] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] created_port_ids = self._update_ports_for_instance( [ 871.419764] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 871.419764] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] with excutils.save_and_reraise_exception(): [ 871.419764] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 871.419764] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] self.force_reraise() [ 871.419764] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 871.419764] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] raise self.value [ 871.419764] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 871.419764] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] updated_port = self._update_port( [ 871.419764] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 871.419764] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] _ensure_no_port_binding_failure(port) [ 871.419764] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 871.419764] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] raise exception.PortBindingFailed(port_id=port['id']) [ 871.420106] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] nova.exception.PortBindingFailed: Binding failed for port 55d081d1-5a3b-4a11-9c16-b73b57c34ea3, please check neutron logs for more information. [ 871.420106] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] [ 871.420106] env[63202]: INFO nova.compute.manager [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] Terminating instance [ 871.420106] env[63202]: DEBUG oslo_concurrency.lockutils [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Acquiring lock "refresh_cache-5d9f1803-e3f8-42a5-b6b7-2642463a079e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.420106] env[63202]: DEBUG oslo_concurrency.lockutils [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Acquired lock "refresh_cache-5d9f1803-e3f8-42a5-b6b7-2642463a079e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.420106] env[63202]: DEBUG nova.network.neutron [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 871.477962] env[63202]: DEBUG nova.network.neutron [-] [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.762912] env[63202]: DEBUG oslo_concurrency.lockutils [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.434s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.763456] env[63202]: DEBUG nova.compute.manager [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: d674db7a-8cf0-46f8-b676-fced0767c105] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 871.766033] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.874s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.768051] env[63202]: INFO nova.compute.claims [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 871.936717] env[63202]: DEBUG nova.network.neutron [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 871.980286] env[63202]: INFO nova.compute.manager [-] [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] Took 1.02 seconds to deallocate network for instance. [ 871.985680] env[63202]: DEBUG nova.compute.claims [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 871.985680] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.051734] env[63202]: DEBUG nova.network.neutron [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.271739] env[63202]: DEBUG nova.compute.utils [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 872.275415] env[63202]: DEBUG nova.compute.manager [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: d674db7a-8cf0-46f8-b676-fced0767c105] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 872.275592] env[63202]: DEBUG nova.network.neutron [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: d674db7a-8cf0-46f8-b676-fced0767c105] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 872.333566] env[63202]: DEBUG nova.policy [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b59e7e62d92e4f5eb0fee41e8f5196d7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f508ec04edc844a19640a8a85f27e5b9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 872.554956] env[63202]: DEBUG oslo_concurrency.lockutils [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Releasing lock "refresh_cache-5d9f1803-e3f8-42a5-b6b7-2642463a079e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.555334] env[63202]: DEBUG nova.compute.manager [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 872.555532] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 872.555826] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1a94f373-e6dd-48f3-9b90-b63b3ae99b51 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.565557] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-301e7e0c-77c2-4571-ae9f-22d3acf2adf7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.589165] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5d9f1803-e3f8-42a5-b6b7-2642463a079e could not be found. [ 872.589463] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 872.589619] env[63202]: INFO nova.compute.manager [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] Took 0.03 seconds to destroy the instance on the hypervisor. [ 872.589903] env[63202]: DEBUG oslo.service.loopingcall [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 872.590144] env[63202]: DEBUG nova.compute.manager [-] [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 872.590238] env[63202]: DEBUG nova.network.neutron [-] [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 872.603201] env[63202]: DEBUG nova.network.neutron [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: d674db7a-8cf0-46f8-b676-fced0767c105] Successfully created port: 8852a0b8-5683-4012-ae86-2dde3afc9dbf {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 872.605955] env[63202]: DEBUG nova.network.neutron [-] [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 872.741749] env[63202]: DEBUG nova.compute.manager [req-0bf27c8e-b68a-4390-8283-08e312c3e0cc req-b279a500-5bf3-4dd8-a7cc-11cc3d49b599 service nova] [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] Received event network-changed-55d081d1-5a3b-4a11-9c16-b73b57c34ea3 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 872.741982] env[63202]: DEBUG nova.compute.manager [req-0bf27c8e-b68a-4390-8283-08e312c3e0cc req-b279a500-5bf3-4dd8-a7cc-11cc3d49b599 service nova] [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] Refreshing instance network info cache due to event network-changed-55d081d1-5a3b-4a11-9c16-b73b57c34ea3. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 872.746327] env[63202]: DEBUG oslo_concurrency.lockutils [req-0bf27c8e-b68a-4390-8283-08e312c3e0cc req-b279a500-5bf3-4dd8-a7cc-11cc3d49b599 service nova] Acquiring lock "refresh_cache-5d9f1803-e3f8-42a5-b6b7-2642463a079e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.746463] env[63202]: DEBUG oslo_concurrency.lockutils [req-0bf27c8e-b68a-4390-8283-08e312c3e0cc req-b279a500-5bf3-4dd8-a7cc-11cc3d49b599 service nova] Acquired lock "refresh_cache-5d9f1803-e3f8-42a5-b6b7-2642463a079e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.746662] env[63202]: DEBUG nova.network.neutron [req-0bf27c8e-b68a-4390-8283-08e312c3e0cc req-b279a500-5bf3-4dd8-a7cc-11cc3d49b599 service nova] [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] Refreshing network info cache for port 55d081d1-5a3b-4a11-9c16-b73b57c34ea3 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 872.775971] env[63202]: DEBUG nova.compute.manager [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: d674db7a-8cf0-46f8-b676-fced0767c105] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 873.030756] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8c3f4e0-d2e4-4d99-a24a-f9f415af4d47 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.045569] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-628dd4b5-b7bf-4745-998d-ee73e116e088 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.078015] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7986d4d6-f327-4091-a5d8-7a42ebb045af {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.085821] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0041af6-79de-40f5-bcc2-092a614367b2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.102074] env[63202]: DEBUG nova.compute.provider_tree [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 873.108691] env[63202]: DEBUG nova.network.neutron [-] [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.266256] env[63202]: DEBUG nova.network.neutron [req-0bf27c8e-b68a-4390-8283-08e312c3e0cc req-b279a500-5bf3-4dd8-a7cc-11cc3d49b599 service nova] [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 873.350293] env[63202]: DEBUG nova.network.neutron [req-0bf27c8e-b68a-4390-8283-08e312c3e0cc req-b279a500-5bf3-4dd8-a7cc-11cc3d49b599 service nova] [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.600569] env[63202]: ERROR nova.compute.manager [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8852a0b8-5683-4012-ae86-2dde3afc9dbf, please check neutron logs for more information. [ 873.600569] env[63202]: ERROR nova.compute.manager Traceback (most recent call last): [ 873.600569] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 873.600569] env[63202]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 873.600569] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 873.600569] env[63202]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 873.600569] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 873.600569] env[63202]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 873.600569] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 873.600569] env[63202]: ERROR nova.compute.manager self.force_reraise() [ 873.600569] env[63202]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 873.600569] env[63202]: ERROR nova.compute.manager raise self.value [ 873.600569] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 873.600569] env[63202]: ERROR nova.compute.manager updated_port = self._update_port( [ 873.600569] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 873.600569] env[63202]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 873.600961] env[63202]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 873.600961] env[63202]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 873.600961] env[63202]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8852a0b8-5683-4012-ae86-2dde3afc9dbf, please check neutron logs for more information. [ 873.600961] env[63202]: ERROR nova.compute.manager [ 873.600961] env[63202]: Traceback (most recent call last): [ 873.600961] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 873.600961] env[63202]: listener.cb(fileno) [ 873.600961] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 873.600961] env[63202]: result = function(*args, **kwargs) [ 873.600961] env[63202]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 873.600961] env[63202]: return func(*args, **kwargs) [ 873.600961] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 873.600961] env[63202]: raise e [ 873.600961] env[63202]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 873.600961] env[63202]: nwinfo = self.network_api.allocate_for_instance( [ 873.600961] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 873.600961] env[63202]: created_port_ids = self._update_ports_for_instance( [ 873.600961] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 873.600961] env[63202]: with excutils.save_and_reraise_exception(): [ 873.600961] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 873.600961] env[63202]: self.force_reraise() [ 873.600961] env[63202]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 873.600961] env[63202]: raise self.value [ 873.600961] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 873.600961] env[63202]: updated_port = self._update_port( [ 873.601577] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 873.601577] env[63202]: _ensure_no_port_binding_failure(port) [ 873.601577] env[63202]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 873.601577] env[63202]: raise exception.PortBindingFailed(port_id=port['id']) [ 873.601577] env[63202]: nova.exception.PortBindingFailed: Binding failed for port 8852a0b8-5683-4012-ae86-2dde3afc9dbf, please check neutron logs for more information. [ 873.601577] env[63202]: Removing descriptor: 16 [ 873.603970] env[63202]: DEBUG nova.scheduler.client.report [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 873.610533] env[63202]: INFO nova.compute.manager [-] [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] Took 1.02 seconds to deallocate network for instance. [ 873.612576] env[63202]: DEBUG nova.compute.claims [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 873.612807] env[63202]: DEBUG oslo_concurrency.lockutils [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.788048] env[63202]: DEBUG nova.compute.manager [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: d674db7a-8cf0-46f8-b676-fced0767c105] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 873.811756] env[63202]: DEBUG nova.virt.hardware [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 873.811997] env[63202]: DEBUG nova.virt.hardware [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 873.812190] env[63202]: DEBUG nova.virt.hardware [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 873.812374] env[63202]: DEBUG nova.virt.hardware [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 873.812516] env[63202]: DEBUG nova.virt.hardware [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 873.812683] env[63202]: DEBUG nova.virt.hardware [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 873.812848] env[63202]: DEBUG nova.virt.hardware [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 873.812998] env[63202]: DEBUG nova.virt.hardware [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 873.813166] env[63202]: DEBUG nova.virt.hardware [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 873.813318] env[63202]: DEBUG nova.virt.hardware [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 873.813486] env[63202]: DEBUG nova.virt.hardware [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 873.814338] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9172686c-5951-441a-86a5-42630aeaa840 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.822155] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dd9926e-29dd-40c4-865a-872a4dd37f2a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.835545] env[63202]: ERROR nova.compute.manager [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: d674db7a-8cf0-46f8-b676-fced0767c105] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8852a0b8-5683-4012-ae86-2dde3afc9dbf, please check neutron logs for more information. [ 873.835545] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] Traceback (most recent call last): [ 873.835545] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 873.835545] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] yield resources [ 873.835545] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 873.835545] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] self.driver.spawn(context, instance, image_meta, [ 873.835545] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 873.835545] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] self._vmops.spawn(context, instance, image_meta, injected_files, [ 873.835545] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 873.835545] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] vm_ref = self.build_virtual_machine(instance, [ 873.835545] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 873.835902] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] vif_infos = vmwarevif.get_vif_info(self._session, [ 873.835902] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 873.835902] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] for vif in network_info: [ 873.835902] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 873.835902] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] return self._sync_wrapper(fn, *args, **kwargs) [ 873.835902] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 873.835902] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] self.wait() [ 873.835902] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 873.835902] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] self[:] = self._gt.wait() [ 873.835902] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 873.835902] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] return self._exit_event.wait() [ 873.835902] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 873.835902] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] current.throw(*self._exc) [ 873.836258] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 873.836258] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] result = function(*args, **kwargs) [ 873.836258] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 873.836258] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] return func(*args, **kwargs) [ 873.836258] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 873.836258] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] raise e [ 873.836258] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 873.836258] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] nwinfo = self.network_api.allocate_for_instance( [ 873.836258] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 873.836258] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] created_port_ids = self._update_ports_for_instance( [ 873.836258] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 873.836258] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] with excutils.save_and_reraise_exception(): [ 873.836258] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 873.836610] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] self.force_reraise() [ 873.836610] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 873.836610] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] raise self.value [ 873.836610] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 873.836610] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] updated_port = self._update_port( [ 873.836610] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 873.836610] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] _ensure_no_port_binding_failure(port) [ 873.836610] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 873.836610] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] raise exception.PortBindingFailed(port_id=port['id']) [ 873.836610] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] nova.exception.PortBindingFailed: Binding failed for port 8852a0b8-5683-4012-ae86-2dde3afc9dbf, please check neutron logs for more information. [ 873.836610] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] [ 873.836610] env[63202]: INFO nova.compute.manager [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: d674db7a-8cf0-46f8-b676-fced0767c105] Terminating instance [ 873.837752] env[63202]: DEBUG oslo_concurrency.lockutils [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "refresh_cache-d674db7a-8cf0-46f8-b676-fced0767c105" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.837908] env[63202]: DEBUG oslo_concurrency.lockutils [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquired lock "refresh_cache-d674db7a-8cf0-46f8-b676-fced0767c105" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.838087] env[63202]: DEBUG nova.network.neutron [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: d674db7a-8cf0-46f8-b676-fced0767c105] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 873.853167] env[63202]: DEBUG oslo_concurrency.lockutils [req-0bf27c8e-b68a-4390-8283-08e312c3e0cc req-b279a500-5bf3-4dd8-a7cc-11cc3d49b599 service nova] Releasing lock "refresh_cache-5d9f1803-e3f8-42a5-b6b7-2642463a079e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.853394] env[63202]: DEBUG nova.compute.manager [req-0bf27c8e-b68a-4390-8283-08e312c3e0cc req-b279a500-5bf3-4dd8-a7cc-11cc3d49b599 service nova] [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] Received event network-vif-deleted-55d081d1-5a3b-4a11-9c16-b73b57c34ea3 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 874.108572] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.342s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.109077] env[63202]: DEBUG nova.compute.manager [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 874.111667] env[63202]: DEBUG oslo_concurrency.lockutils [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.474s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.113682] env[63202]: INFO nova.compute.claims [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 874.358056] env[63202]: DEBUG nova.network.neutron [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: d674db7a-8cf0-46f8-b676-fced0767c105] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 874.438304] env[63202]: DEBUG nova.network.neutron [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: d674db7a-8cf0-46f8-b676-fced0767c105] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.618315] env[63202]: DEBUG nova.compute.utils [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 874.624071] env[63202]: DEBUG nova.compute.manager [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 874.624293] env[63202]: DEBUG nova.network.neutron [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 874.666731] env[63202]: DEBUG nova.policy [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0db0311686c44dfba9f41d0d50f95590', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9ac1eb57753e4323987128674d291627', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 874.766565] env[63202]: DEBUG nova.compute.manager [req-8b228a85-a6a8-412e-bbc3-100ef74c102f req-2ed9890c-5ef0-4e03-9d2b-f661c443f18f service nova] [instance: d674db7a-8cf0-46f8-b676-fced0767c105] Received event network-changed-8852a0b8-5683-4012-ae86-2dde3afc9dbf {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 874.766766] env[63202]: DEBUG nova.compute.manager [req-8b228a85-a6a8-412e-bbc3-100ef74c102f req-2ed9890c-5ef0-4e03-9d2b-f661c443f18f service nova] [instance: d674db7a-8cf0-46f8-b676-fced0767c105] Refreshing instance network info cache due to event network-changed-8852a0b8-5683-4012-ae86-2dde3afc9dbf. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 874.766953] env[63202]: DEBUG oslo_concurrency.lockutils [req-8b228a85-a6a8-412e-bbc3-100ef74c102f req-2ed9890c-5ef0-4e03-9d2b-f661c443f18f service nova] Acquiring lock "refresh_cache-d674db7a-8cf0-46f8-b676-fced0767c105" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.947109] env[63202]: DEBUG oslo_concurrency.lockutils [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Releasing lock "refresh_cache-d674db7a-8cf0-46f8-b676-fced0767c105" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.947556] env[63202]: DEBUG nova.compute.manager [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: d674db7a-8cf0-46f8-b676-fced0767c105] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 874.947756] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: d674db7a-8cf0-46f8-b676-fced0767c105] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 874.948084] env[63202]: DEBUG oslo_concurrency.lockutils [req-8b228a85-a6a8-412e-bbc3-100ef74c102f req-2ed9890c-5ef0-4e03-9d2b-f661c443f18f service nova] Acquired lock "refresh_cache-d674db7a-8cf0-46f8-b676-fced0767c105" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.948258] env[63202]: DEBUG nova.network.neutron [req-8b228a85-a6a8-412e-bbc3-100ef74c102f req-2ed9890c-5ef0-4e03-9d2b-f661c443f18f service nova] [instance: d674db7a-8cf0-46f8-b676-fced0767c105] Refreshing network info cache for port 8852a0b8-5683-4012-ae86-2dde3afc9dbf {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 874.949582] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c6d6c90f-13ca-4bea-ad56-42795eb6056f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.961497] env[63202]: DEBUG nova.network.neutron [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Successfully created port: f09247c0-c4a4-4583-95b6-54de5ec7b011 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 874.969344] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45fc4e01-7cfa-4cf5-a9ba-ed65835d06f4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.997369] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: d674db7a-8cf0-46f8-b676-fced0767c105] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d674db7a-8cf0-46f8-b676-fced0767c105 could not be found. [ 874.997555] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: d674db7a-8cf0-46f8-b676-fced0767c105] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 874.997727] env[63202]: INFO nova.compute.manager [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: d674db7a-8cf0-46f8-b676-fced0767c105] Took 0.05 seconds to destroy the instance on the hypervisor. [ 874.997961] env[63202]: DEBUG oslo.service.loopingcall [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 874.998194] env[63202]: DEBUG nova.compute.manager [-] [instance: d674db7a-8cf0-46f8-b676-fced0767c105] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 874.998284] env[63202]: DEBUG nova.network.neutron [-] [instance: d674db7a-8cf0-46f8-b676-fced0767c105] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 875.016932] env[63202]: DEBUG nova.network.neutron [-] [instance: d674db7a-8cf0-46f8-b676-fced0767c105] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 875.130758] env[63202]: DEBUG nova.compute.manager [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 875.431890] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0908437-f0ee-45f9-816e-233cf40eb799 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.439870] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7607fdd1-6ad6-4565-bc6c-a5a56c106306 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.469858] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-161c2e6f-0152-4048-bb43-f9f5f219acaf {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.476711] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aef9bded-e3e7-4cd7-999c-838dff009779 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.489296] env[63202]: DEBUG nova.compute.provider_tree [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 875.490814] env[63202]: DEBUG nova.network.neutron [req-8b228a85-a6a8-412e-bbc3-100ef74c102f req-2ed9890c-5ef0-4e03-9d2b-f661c443f18f service nova] [instance: d674db7a-8cf0-46f8-b676-fced0767c105] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 875.522196] env[63202]: DEBUG nova.network.neutron [-] [instance: d674db7a-8cf0-46f8-b676-fced0767c105] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.568936] env[63202]: DEBUG nova.network.neutron [req-8b228a85-a6a8-412e-bbc3-100ef74c102f req-2ed9890c-5ef0-4e03-9d2b-f661c443f18f service nova] [instance: d674db7a-8cf0-46f8-b676-fced0767c105] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.993983] env[63202]: DEBUG nova.scheduler.client.report [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 876.025156] env[63202]: INFO nova.compute.manager [-] [instance: d674db7a-8cf0-46f8-b676-fced0767c105] Took 1.03 seconds to deallocate network for instance. [ 876.028173] env[63202]: DEBUG nova.compute.claims [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: d674db7a-8cf0-46f8-b676-fced0767c105] Aborting claim: {{(pid=63202) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 876.028173] env[63202]: DEBUG oslo_concurrency.lockutils [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.071764] env[63202]: DEBUG oslo_concurrency.lockutils [req-8b228a85-a6a8-412e-bbc3-100ef74c102f req-2ed9890c-5ef0-4e03-9d2b-f661c443f18f service nova] Releasing lock "refresh_cache-d674db7a-8cf0-46f8-b676-fced0767c105" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.072013] env[63202]: DEBUG nova.compute.manager [req-8b228a85-a6a8-412e-bbc3-100ef74c102f req-2ed9890c-5ef0-4e03-9d2b-f661c443f18f service nova] [instance: d674db7a-8cf0-46f8-b676-fced0767c105] Received event network-vif-deleted-8852a0b8-5683-4012-ae86-2dde3afc9dbf {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 876.143376] env[63202]: DEBUG nova.compute.manager [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 876.167201] env[63202]: DEBUG nova.virt.hardware [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 876.167460] env[63202]: DEBUG nova.virt.hardware [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 876.167611] env[63202]: DEBUG nova.virt.hardware [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 876.167813] env[63202]: DEBUG nova.virt.hardware [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 876.167962] env[63202]: DEBUG nova.virt.hardware [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 876.168117] env[63202]: DEBUG nova.virt.hardware [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 876.168320] env[63202]: DEBUG nova.virt.hardware [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 876.168471] env[63202]: DEBUG nova.virt.hardware [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 876.168627] env[63202]: DEBUG nova.virt.hardware [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 876.168794] env[63202]: DEBUG nova.virt.hardware [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 876.168960] env[63202]: DEBUG nova.virt.hardware [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 876.169906] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d3e6e24-e8ba-42c0-8f8f-323bf5acbf7f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.178138] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70b60064-1a6b-47d2-8433-a60a2069bd74 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.498935] env[63202]: DEBUG oslo_concurrency.lockutils [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.387s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.499492] env[63202]: DEBUG nova.compute.manager [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 876.502631] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.037s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.506484] env[63202]: INFO nova.compute.claims [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] [instance: a800fa5d-46a1-4a96-8ea2-26e34b869aa6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 876.830624] env[63202]: DEBUG nova.compute.manager [req-5fb0b5f0-73f4-47d2-ad60-7163cfc1c9f8 req-8604acbe-051e-4341-aacb-f9c3d2f97e39 service nova] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Received event network-vif-plugged-f09247c0-c4a4-4583-95b6-54de5ec7b011 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 876.830907] env[63202]: DEBUG oslo_concurrency.lockutils [req-5fb0b5f0-73f4-47d2-ad60-7163cfc1c9f8 req-8604acbe-051e-4341-aacb-f9c3d2f97e39 service nova] Acquiring lock "cedc3a06-2123-4c5e-a6c3-599a3efc3c65-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.831150] env[63202]: DEBUG oslo_concurrency.lockutils [req-5fb0b5f0-73f4-47d2-ad60-7163cfc1c9f8 req-8604acbe-051e-4341-aacb-f9c3d2f97e39 service nova] Lock "cedc3a06-2123-4c5e-a6c3-599a3efc3c65-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.831321] env[63202]: DEBUG oslo_concurrency.lockutils [req-5fb0b5f0-73f4-47d2-ad60-7163cfc1c9f8 req-8604acbe-051e-4341-aacb-f9c3d2f97e39 service nova] Lock "cedc3a06-2123-4c5e-a6c3-599a3efc3c65-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.831489] env[63202]: DEBUG nova.compute.manager [req-5fb0b5f0-73f4-47d2-ad60-7163cfc1c9f8 req-8604acbe-051e-4341-aacb-f9c3d2f97e39 service nova] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] No waiting events found dispatching network-vif-plugged-f09247c0-c4a4-4583-95b6-54de5ec7b011 {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 876.831653] env[63202]: WARNING nova.compute.manager [req-5fb0b5f0-73f4-47d2-ad60-7163cfc1c9f8 req-8604acbe-051e-4341-aacb-f9c3d2f97e39 service nova] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Received unexpected event network-vif-plugged-f09247c0-c4a4-4583-95b6-54de5ec7b011 for instance with vm_state building and task_state spawning. [ 876.896850] env[63202]: DEBUG nova.network.neutron [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Successfully updated port: f09247c0-c4a4-4583-95b6-54de5ec7b011 {{(pid=63202) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 877.008653] env[63202]: DEBUG nova.compute.utils [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 877.012089] env[63202]: DEBUG nova.compute.manager [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 877.012265] env[63202]: DEBUG nova.network.neutron [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 877.052701] env[63202]: DEBUG nova.policy [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7e11242d52d14577847d17d3a6441202', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a9e939cdf4a94a8aa0107a5761771c2c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 877.363669] env[63202]: DEBUG nova.network.neutron [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Successfully created port: e5621970-d0a8-4803-94ef-f6d577bdf6d2 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 877.399431] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Acquiring lock "refresh_cache-cedc3a06-2123-4c5e-a6c3-599a3efc3c65" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.399592] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Acquired lock "refresh_cache-cedc3a06-2123-4c5e-a6c3-599a3efc3c65" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.399771] env[63202]: DEBUG nova.network.neutron [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 877.513314] env[63202]: DEBUG nova.compute.manager [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 877.785317] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94b2907c-40c6-4127-8cb8-5ee0afadab7d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.792901] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3269b7f8-3712-47da-90bd-86dfea21c6df {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.822323] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e3178a7-782d-49a6-841a-3c3087e6802b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.828821] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cd6dcbf-6cbb-4159-b019-b291a1498369 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.841190] env[63202]: DEBUG nova.compute.provider_tree [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 877.930156] env[63202]: DEBUG nova.network.neutron [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 878.110591] env[63202]: DEBUG nova.network.neutron [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Updating instance_info_cache with network_info: [{"id": "f09247c0-c4a4-4583-95b6-54de5ec7b011", "address": "fa:16:3e:9e:74:85", "network": {"id": "256de701-9fea-4097-aebc-8f0a7acff6ee", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1676107181-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ac1eb57753e4323987128674d291627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6a6f7bb-6106-4cfd-9aef-b85628d0cefa", "external-id": "nsx-vlan-transportzone-194", "segmentation_id": 194, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf09247c0-c4", "ovs_interfaceid": "f09247c0-c4a4-4583-95b6-54de5ec7b011", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.345781] env[63202]: DEBUG nova.scheduler.client.report [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 878.532254] env[63202]: DEBUG nova.compute.manager [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 878.556719] env[63202]: DEBUG nova.virt.hardware [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 878.556959] env[63202]: DEBUG nova.virt.hardware [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 878.557129] env[63202]: DEBUG nova.virt.hardware [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 878.557309] env[63202]: DEBUG nova.virt.hardware [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 878.557450] env[63202]: DEBUG nova.virt.hardware [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 878.557595] env[63202]: DEBUG nova.virt.hardware [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 878.557802] env[63202]: DEBUG nova.virt.hardware [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 878.557957] env[63202]: DEBUG nova.virt.hardware [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 878.558133] env[63202]: DEBUG nova.virt.hardware [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 878.558292] env[63202]: DEBUG nova.virt.hardware [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 878.558459] env[63202]: DEBUG nova.virt.hardware [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 878.559312] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fc51a6d-7f7f-4509-8b92-d4d15d281c3d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.567123] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58f89be7-a795-4906-b703-218b771fa8f0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.613660] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Releasing lock "refresh_cache-cedc3a06-2123-4c5e-a6c3-599a3efc3c65" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.613974] env[63202]: DEBUG nova.compute.manager [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Instance network_info: |[{"id": "f09247c0-c4a4-4583-95b6-54de5ec7b011", "address": "fa:16:3e:9e:74:85", "network": {"id": "256de701-9fea-4097-aebc-8f0a7acff6ee", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1676107181-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ac1eb57753e4323987128674d291627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6a6f7bb-6106-4cfd-9aef-b85628d0cefa", "external-id": "nsx-vlan-transportzone-194", "segmentation_id": 194, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf09247c0-c4", "ovs_interfaceid": "f09247c0-c4a4-4583-95b6-54de5ec7b011", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 878.614473] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9e:74:85', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd6a6f7bb-6106-4cfd-9aef-b85628d0cefa', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f09247c0-c4a4-4583-95b6-54de5ec7b011', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 878.622324] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Creating folder: Project (9ac1eb57753e4323987128674d291627). Parent ref: group-v294090. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 878.622903] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e0c46efd-5b86-4ed5-85b1-c5da66915fcd {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.633336] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Created folder: Project (9ac1eb57753e4323987128674d291627) in parent group-v294090. [ 878.633506] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Creating folder: Instances. Parent ref: group-v294112. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 878.633717] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fddbe216-ef04-47f9-93a3-d52aed916c9e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.641391] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Created folder: Instances in parent group-v294112. [ 878.641600] env[63202]: DEBUG oslo.service.loopingcall [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 878.641773] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 878.641954] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-61f558cd-6e12-4a84-b0bf-8bf3e8293122 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.658789] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 878.658789] env[63202]: value = "task-1385379" [ 878.658789] env[63202]: _type = "Task" [ 878.658789] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.667288] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385379, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.849906] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.347s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.850469] env[63202]: DEBUG nova.compute.manager [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] [instance: a800fa5d-46a1-4a96-8ea2-26e34b869aa6] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 878.853152] env[63202]: DEBUG oslo_concurrency.lockutils [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.503s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.854523] env[63202]: INFO nova.compute.claims [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 878.859290] env[63202]: DEBUG nova.compute.manager [req-d5ac31da-3dc9-4fe4-8e00-0b48a0800494 req-7ab4fc7e-835b-4dcf-83a7-aba0492459cc service nova] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Received event network-changed-f09247c0-c4a4-4583-95b6-54de5ec7b011 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 878.859501] env[63202]: DEBUG nova.compute.manager [req-d5ac31da-3dc9-4fe4-8e00-0b48a0800494 req-7ab4fc7e-835b-4dcf-83a7-aba0492459cc service nova] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Refreshing instance network info cache due to event network-changed-f09247c0-c4a4-4583-95b6-54de5ec7b011. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 878.859747] env[63202]: DEBUG oslo_concurrency.lockutils [req-d5ac31da-3dc9-4fe4-8e00-0b48a0800494 req-7ab4fc7e-835b-4dcf-83a7-aba0492459cc service nova] Acquiring lock "refresh_cache-cedc3a06-2123-4c5e-a6c3-599a3efc3c65" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.859894] env[63202]: DEBUG oslo_concurrency.lockutils [req-d5ac31da-3dc9-4fe4-8e00-0b48a0800494 req-7ab4fc7e-835b-4dcf-83a7-aba0492459cc service nova] Acquired lock "refresh_cache-cedc3a06-2123-4c5e-a6c3-599a3efc3c65" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.860095] env[63202]: DEBUG nova.network.neutron [req-d5ac31da-3dc9-4fe4-8e00-0b48a0800494 req-7ab4fc7e-835b-4dcf-83a7-aba0492459cc service nova] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Refreshing network info cache for port f09247c0-c4a4-4583-95b6-54de5ec7b011 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 879.015393] env[63202]: DEBUG nova.network.neutron [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Successfully updated port: e5621970-d0a8-4803-94ef-f6d577bdf6d2 {{(pid=63202) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 879.169129] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385379, 'name': CreateVM_Task} progress is 99%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.253900] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 879.360064] env[63202]: DEBUG nova.compute.utils [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 879.361573] env[63202]: DEBUG nova.compute.manager [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] [instance: a800fa5d-46a1-4a96-8ea2-26e34b869aa6] Not allocating networking since 'none' was specified. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 879.518176] env[63202]: DEBUG oslo_concurrency.lockutils [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Acquiring lock "refresh_cache-0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.518293] env[63202]: DEBUG oslo_concurrency.lockutils [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Acquired lock "refresh_cache-0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.518427] env[63202]: DEBUG nova.network.neutron [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 879.669657] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385379, 'name': CreateVM_Task} progress is 99%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.684179] env[63202]: DEBUG nova.network.neutron [req-d5ac31da-3dc9-4fe4-8e00-0b48a0800494 req-7ab4fc7e-835b-4dcf-83a7-aba0492459cc service nova] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Updated VIF entry in instance network info cache for port f09247c0-c4a4-4583-95b6-54de5ec7b011. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 879.684528] env[63202]: DEBUG nova.network.neutron [req-d5ac31da-3dc9-4fe4-8e00-0b48a0800494 req-7ab4fc7e-835b-4dcf-83a7-aba0492459cc service nova] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Updating instance_info_cache with network_info: [{"id": "f09247c0-c4a4-4583-95b6-54de5ec7b011", "address": "fa:16:3e:9e:74:85", "network": {"id": "256de701-9fea-4097-aebc-8f0a7acff6ee", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1676107181-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ac1eb57753e4323987128674d291627", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6a6f7bb-6106-4cfd-9aef-b85628d0cefa", "external-id": "nsx-vlan-transportzone-194", "segmentation_id": 194, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf09247c0-c4", "ovs_interfaceid": "f09247c0-c4a4-4583-95b6-54de5ec7b011", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.865134] env[63202]: DEBUG nova.compute.manager [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] [instance: a800fa5d-46a1-4a96-8ea2-26e34b869aa6] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 880.047375] env[63202]: DEBUG nova.network.neutron [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 880.111453] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bfeecd7-1be0-45bd-a10f-a58f4080e8c5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.118468] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bdf1674-683b-4315-a2ca-9e1739a54fd2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.149167] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aec77bfd-ad20-48d1-a410-c78718602fa7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.157052] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ada248d5-9bee-45ab-829c-d22a21676d73 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.175224] env[63202]: DEBUG nova.compute.provider_tree [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 880.180167] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385379, 'name': CreateVM_Task, 'duration_secs': 1.294752} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.180306] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 880.186960] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.187185] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.187729] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 880.187994] env[63202]: DEBUG oslo_concurrency.lockutils [req-d5ac31da-3dc9-4fe4-8e00-0b48a0800494 req-7ab4fc7e-835b-4dcf-83a7-aba0492459cc service nova] Releasing lock "refresh_cache-cedc3a06-2123-4c5e-a6c3-599a3efc3c65" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.188551] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d56c2456-5bb1-42f7-940a-02156d481d5e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.193779] env[63202]: DEBUG oslo_vmware.api [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Waiting for the task: (returnval){ [ 880.193779] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52f9f0c8-69f7-0d48-5ee4-e9acc0d9b1f6" [ 880.193779] env[63202]: _type = "Task" [ 880.193779] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.201428] env[63202]: DEBUG oslo_vmware.api [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52f9f0c8-69f7-0d48-5ee4-e9acc0d9b1f6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.220625] env[63202]: DEBUG nova.network.neutron [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Updating instance_info_cache with network_info: [{"id": "e5621970-d0a8-4803-94ef-f6d577bdf6d2", "address": "fa:16:3e:00:ba:72", "network": {"id": "345538f4-e791-4fc6-9719-237c11f1382a", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-449651008-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a9e939cdf4a94a8aa0107a5761771c2c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa09e855-8af1-419b-b78d-8ffcc94b1bfb", "external-id": "nsx-vlan-transportzone-901", "segmentation_id": 901, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape5621970-d0", "ovs_interfaceid": "e5621970-d0a8-4803-94ef-f6d577bdf6d2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.253517] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 880.253706] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 880.681360] env[63202]: DEBUG nova.scheduler.client.report [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 880.705055] env[63202]: DEBUG oslo_vmware.api [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52f9f0c8-69f7-0d48-5ee4-e9acc0d9b1f6, 'name': SearchDatastore_Task, 'duration_secs': 0.009663} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.705344] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.705565] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 880.705787] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.705931] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.706115] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 880.706357] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8c82eb6a-0314-40d8-b537-b9b1919c00e8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.714029] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 880.714211] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 880.714897] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-23d016d7-81db-4ce0-925f-d7bd3dde797b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.719862] env[63202]: DEBUG oslo_vmware.api [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Waiting for the task: (returnval){ [ 880.719862] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52ca6471-6e39-bb1b-f026-f319f3d11b79" [ 880.719862] env[63202]: _type = "Task" [ 880.719862] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.722947] env[63202]: DEBUG oslo_concurrency.lockutils [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Releasing lock "refresh_cache-0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.723214] env[63202]: DEBUG nova.compute.manager [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Instance network_info: |[{"id": "e5621970-d0a8-4803-94ef-f6d577bdf6d2", "address": "fa:16:3e:00:ba:72", "network": {"id": "345538f4-e791-4fc6-9719-237c11f1382a", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-449651008-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a9e939cdf4a94a8aa0107a5761771c2c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa09e855-8af1-419b-b78d-8ffcc94b1bfb", "external-id": "nsx-vlan-transportzone-901", "segmentation_id": 901, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape5621970-d0", "ovs_interfaceid": "e5621970-d0a8-4803-94ef-f6d577bdf6d2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 880.723605] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:00:ba:72', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'aa09e855-8af1-419b-b78d-8ffcc94b1bfb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e5621970-d0a8-4803-94ef-f6d577bdf6d2', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 880.731515] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Creating folder: Project (a9e939cdf4a94a8aa0107a5761771c2c). Parent ref: group-v294090. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 880.732094] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6a86070f-4c39-4779-ac66-f4054a551f7c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.737218] env[63202]: DEBUG oslo_vmware.api [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52ca6471-6e39-bb1b-f026-f319f3d11b79, 'name': SearchDatastore_Task, 'duration_secs': 0.010728} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.738205] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6a4016c6-b1d7-4ca6-b191-2e37681a601e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.742857] env[63202]: DEBUG oslo_vmware.api [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Waiting for the task: (returnval){ [ 880.742857] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]520df2fe-d519-3eb5-55fc-d796a2127c28" [ 880.742857] env[63202]: _type = "Task" [ 880.742857] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.746564] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Created folder: Project (a9e939cdf4a94a8aa0107a5761771c2c) in parent group-v294090. [ 880.746739] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Creating folder: Instances. Parent ref: group-v294115. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 880.747211] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6c960727-aa3b-4f77-a60a-ae70ae418a57 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.751323] env[63202]: DEBUG oslo_vmware.api [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]520df2fe-d519-3eb5-55fc-d796a2127c28, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.756647] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Created folder: Instances in parent group-v294115. [ 880.756864] env[63202]: DEBUG oslo.service.loopingcall [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 880.757070] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 880.757256] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f3f1e798-e0e8-464e-8b45-1a92c1e57e56 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.775085] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 880.775085] env[63202]: value = "task-1385382" [ 880.775085] env[63202]: _type = "Task" [ 880.775085] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.783302] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385382, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.877318] env[63202]: DEBUG nova.compute.manager [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] [instance: a800fa5d-46a1-4a96-8ea2-26e34b869aa6] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 880.884761] env[63202]: DEBUG nova.compute.manager [req-d05e198b-dd82-47d6-98db-d0e8fb7d06e7 req-4c4884a0-e187-4286-92eb-d4f6fb15d839 service nova] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Received event network-vif-plugged-e5621970-d0a8-4803-94ef-f6d577bdf6d2 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 880.884990] env[63202]: DEBUG oslo_concurrency.lockutils [req-d05e198b-dd82-47d6-98db-d0e8fb7d06e7 req-4c4884a0-e187-4286-92eb-d4f6fb15d839 service nova] Acquiring lock "0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.887060] env[63202]: DEBUG oslo_concurrency.lockutils [req-d05e198b-dd82-47d6-98db-d0e8fb7d06e7 req-4c4884a0-e187-4286-92eb-d4f6fb15d839 service nova] Lock "0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.002s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.887309] env[63202]: DEBUG oslo_concurrency.lockutils [req-d05e198b-dd82-47d6-98db-d0e8fb7d06e7 req-4c4884a0-e187-4286-92eb-d4f6fb15d839 service nova] Lock "0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.887553] env[63202]: DEBUG nova.compute.manager [req-d05e198b-dd82-47d6-98db-d0e8fb7d06e7 req-4c4884a0-e187-4286-92eb-d4f6fb15d839 service nova] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] No waiting events found dispatching network-vif-plugged-e5621970-d0a8-4803-94ef-f6d577bdf6d2 {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 880.887856] env[63202]: WARNING nova.compute.manager [req-d05e198b-dd82-47d6-98db-d0e8fb7d06e7 req-4c4884a0-e187-4286-92eb-d4f6fb15d839 service nova] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Received unexpected event network-vif-plugged-e5621970-d0a8-4803-94ef-f6d577bdf6d2 for instance with vm_state building and task_state spawning. [ 880.888065] env[63202]: DEBUG nova.compute.manager [req-d05e198b-dd82-47d6-98db-d0e8fb7d06e7 req-4c4884a0-e187-4286-92eb-d4f6fb15d839 service nova] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Received event network-changed-e5621970-d0a8-4803-94ef-f6d577bdf6d2 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 880.888305] env[63202]: DEBUG nova.compute.manager [req-d05e198b-dd82-47d6-98db-d0e8fb7d06e7 req-4c4884a0-e187-4286-92eb-d4f6fb15d839 service nova] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Refreshing instance network info cache due to event network-changed-e5621970-d0a8-4803-94ef-f6d577bdf6d2. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 880.888559] env[63202]: DEBUG oslo_concurrency.lockutils [req-d05e198b-dd82-47d6-98db-d0e8fb7d06e7 req-4c4884a0-e187-4286-92eb-d4f6fb15d839 service nova] Acquiring lock "refresh_cache-0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.888730] env[63202]: DEBUG oslo_concurrency.lockutils [req-d05e198b-dd82-47d6-98db-d0e8fb7d06e7 req-4c4884a0-e187-4286-92eb-d4f6fb15d839 service nova] Acquired lock "refresh_cache-0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.888980] env[63202]: DEBUG nova.network.neutron [req-d05e198b-dd82-47d6-98db-d0e8fb7d06e7 req-4c4884a0-e187-4286-92eb-d4f6fb15d839 service nova] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Refreshing network info cache for port e5621970-d0a8-4803-94ef-f6d577bdf6d2 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 880.904782] env[63202]: DEBUG nova.virt.hardware [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 880.905019] env[63202]: DEBUG nova.virt.hardware [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 880.905180] env[63202]: DEBUG nova.virt.hardware [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 880.905359] env[63202]: DEBUG nova.virt.hardware [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 880.905500] env[63202]: DEBUG nova.virt.hardware [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 880.905642] env[63202]: DEBUG nova.virt.hardware [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 880.905847] env[63202]: DEBUG nova.virt.hardware [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 880.906058] env[63202]: DEBUG nova.virt.hardware [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 880.906274] env[63202]: DEBUG nova.virt.hardware [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 880.906441] env[63202]: DEBUG nova.virt.hardware [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 880.906609] env[63202]: DEBUG nova.virt.hardware [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 880.907909] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33bffe81-0e82-40d8-860a-ef248f4d039d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.915840] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe25f650-a290-4511-9c04-278d2bdc4fa0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.929206] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] [instance: a800fa5d-46a1-4a96-8ea2-26e34b869aa6] Instance VIF info [] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 880.934774] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Creating folder: Project (0313d80afcc84597a24afb8b185f6ffc). Parent ref: group-v294090. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 880.935072] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-aed86250-2ed0-4503-977b-7477140db779 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.944375] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Created folder: Project (0313d80afcc84597a24afb8b185f6ffc) in parent group-v294090. [ 880.944558] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Creating folder: Instances. Parent ref: group-v294118. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 880.945136] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-61a45b48-ce9d-4bdd-97db-6e3e05aa00b7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.952869] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Created folder: Instances in parent group-v294118. [ 880.953252] env[63202]: DEBUG oslo.service.loopingcall [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 880.953452] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a800fa5d-46a1-4a96-8ea2-26e34b869aa6] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 880.953641] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b6c504ee-6ea5-47a7-aec5-744a13161e8e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.972261] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 880.972261] env[63202]: value = "task-1385385" [ 880.972261] env[63202]: _type = "Task" [ 880.972261] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.980535] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385385, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.186699] env[63202]: DEBUG oslo_concurrency.lockutils [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.333s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.186967] env[63202]: DEBUG nova.compute.manager [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 881.189755] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.940s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.191375] env[63202]: INFO nova.compute.claims [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 881.248048] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 881.252082] env[63202]: DEBUG oslo_vmware.api [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]520df2fe-d519-3eb5-55fc-d796a2127c28, 'name': SearchDatastore_Task, 'duration_secs': 0.008431} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.252364] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.252617] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] cedc3a06-2123-4c5e-a6c3-599a3efc3c65/cedc3a06-2123-4c5e-a6c3-599a3efc3c65.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 881.252864] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c48d292a-c1fd-435e-b28a-d60742be20c6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.259654] env[63202]: DEBUG oslo_vmware.api [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Waiting for the task: (returnval){ [ 881.259654] env[63202]: value = "task-1385386" [ 881.259654] env[63202]: _type = "Task" [ 881.259654] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.267028] env[63202]: DEBUG oslo_vmware.api [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Task: {'id': task-1385386, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.283146] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385382, 'name': CreateVM_Task, 'duration_secs': 0.299839} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.283296] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 881.283910] env[63202]: DEBUG oslo_concurrency.lockutils [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.284082] env[63202]: DEBUG oslo_concurrency.lockutils [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.284384] env[63202]: DEBUG oslo_concurrency.lockutils [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 881.284607] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-de694574-c2dd-441b-99f9-81b806795b77 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.288320] env[63202]: DEBUG oslo_vmware.api [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Waiting for the task: (returnval){ [ 881.288320] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52471ef8-aa1c-3418-7bde-518755d58dac" [ 881.288320] env[63202]: _type = "Task" [ 881.288320] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.295264] env[63202]: DEBUG oslo_vmware.api [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52471ef8-aa1c-3418-7bde-518755d58dac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.483320] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385385, 'name': CreateVM_Task, 'duration_secs': 0.253536} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.483530] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a800fa5d-46a1-4a96-8ea2-26e34b869aa6] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 881.484088] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.620608] env[63202]: DEBUG nova.network.neutron [req-d05e198b-dd82-47d6-98db-d0e8fb7d06e7 req-4c4884a0-e187-4286-92eb-d4f6fb15d839 service nova] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Updated VIF entry in instance network info cache for port e5621970-d0a8-4803-94ef-f6d577bdf6d2. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 881.621011] env[63202]: DEBUG nova.network.neutron [req-d05e198b-dd82-47d6-98db-d0e8fb7d06e7 req-4c4884a0-e187-4286-92eb-d4f6fb15d839 service nova] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Updating instance_info_cache with network_info: [{"id": "e5621970-d0a8-4803-94ef-f6d577bdf6d2", "address": "fa:16:3e:00:ba:72", "network": {"id": "345538f4-e791-4fc6-9719-237c11f1382a", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-449651008-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a9e939cdf4a94a8aa0107a5761771c2c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa09e855-8af1-419b-b78d-8ffcc94b1bfb", "external-id": "nsx-vlan-transportzone-901", "segmentation_id": 901, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape5621970-d0", "ovs_interfaceid": "e5621970-d0a8-4803-94ef-f6d577bdf6d2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.696159] env[63202]: DEBUG nova.compute.utils [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 881.700244] env[63202]: DEBUG nova.compute.manager [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 881.700445] env[63202]: DEBUG nova.network.neutron [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 881.749932] env[63202]: DEBUG nova.policy [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ec07eaa606564f499b1e735409881c8d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e2926b8bd92644deb614a386f00be6c8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 881.771033] env[63202]: DEBUG oslo_vmware.api [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Task: {'id': task-1385386, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.797270] env[63202]: DEBUG oslo_vmware.api [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52471ef8-aa1c-3418-7bde-518755d58dac, 'name': SearchDatastore_Task, 'duration_secs': 0.015649} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.797541] env[63202]: DEBUG oslo_concurrency.lockutils [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.797813] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 881.798099] env[63202]: DEBUG oslo_concurrency.lockutils [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.798250] env[63202]: DEBUG oslo_concurrency.lockutils [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.798424] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 881.798686] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.798979] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 881.799228] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d13778c9-d806-4c4a-8999-f629e5ab2b13 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.800897] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-00fa56c2-c0e5-4b3d-86ce-ab11f435d715 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.805368] env[63202]: DEBUG oslo_vmware.api [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Waiting for the task: (returnval){ [ 881.805368] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]5204cb1a-8287-8cde-1c24-656f8fcd9ebd" [ 881.805368] env[63202]: _type = "Task" [ 881.805368] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.809551] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 881.809748] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 881.810669] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7380127f-64fb-4544-927a-fd470e026e47 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.815211] env[63202]: DEBUG oslo_vmware.api [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]5204cb1a-8287-8cde-1c24-656f8fcd9ebd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.817777] env[63202]: DEBUG oslo_vmware.api [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Waiting for the task: (returnval){ [ 881.817777] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]526adb40-04e2-250e-694c-9eb03b773cf6" [ 881.817777] env[63202]: _type = "Task" [ 881.817777] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.824859] env[63202]: DEBUG oslo_vmware.api [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]526adb40-04e2-250e-694c-9eb03b773cf6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.020591] env[63202]: DEBUG nova.network.neutron [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Successfully created port: d12e501a-1681-4490-84f1-f48efcf89671 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 882.123390] env[63202]: DEBUG oslo_concurrency.lockutils [req-d05e198b-dd82-47d6-98db-d0e8fb7d06e7 req-4c4884a0-e187-4286-92eb-d4f6fb15d839 service nova] Releasing lock "refresh_cache-0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.203197] env[63202]: DEBUG nova.compute.manager [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 882.255086] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 882.255244] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63202) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 882.270971] env[63202]: DEBUG oslo_vmware.api [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Task: {'id': task-1385386, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.318362] env[63202]: DEBUG oslo_vmware.api [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]5204cb1a-8287-8cde-1c24-656f8fcd9ebd, 'name': SearchDatastore_Task, 'duration_secs': 0.008545} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.318362] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.318534] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] [instance: a800fa5d-46a1-4a96-8ea2-26e34b869aa6] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 882.318743] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.329534] env[63202]: DEBUG oslo_vmware.api [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]526adb40-04e2-250e-694c-9eb03b773cf6, 'name': SearchDatastore_Task, 'duration_secs': 0.010965} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.330088] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b964e4e1-8f20-4d12-845f-e3a8801109ed {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.335376] env[63202]: DEBUG oslo_vmware.api [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Waiting for the task: (returnval){ [ 882.335376] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52d7fe9b-2f5f-ac62-ef68-a799ed33d02f" [ 882.335376] env[63202]: _type = "Task" [ 882.335376] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.346083] env[63202]: DEBUG oslo_vmware.api [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52d7fe9b-2f5f-ac62-ef68-a799ed33d02f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.500523] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba2c0098-8cd0-4f28-80f0-431bde4dea66 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.508086] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef86b94b-a9c2-49f1-b03d-aa172a2c4aca {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.540056] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-475f7467-1cb6-41f5-932d-a76544923539 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.547808] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec5eaf46-0cdd-4403-b46b-b9ac8e0e61fc {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.561662] env[63202]: DEBUG nova.compute.provider_tree [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 882.771503] env[63202]: DEBUG oslo_vmware.api [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Task: {'id': task-1385386, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.345479} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.771765] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] cedc3a06-2123-4c5e-a6c3-599a3efc3c65/cedc3a06-2123-4c5e-a6c3-599a3efc3c65.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 882.771957] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 882.772235] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-01141432-abe5-4c90-a9b9-b27a05bb115e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.780470] env[63202]: DEBUG oslo_vmware.api [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Waiting for the task: (returnval){ [ 882.780470] env[63202]: value = "task-1385387" [ 882.780470] env[63202]: _type = "Task" [ 882.780470] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.788486] env[63202]: DEBUG oslo_vmware.api [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Task: {'id': task-1385387, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.845297] env[63202]: DEBUG oslo_vmware.api [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52d7fe9b-2f5f-ac62-ef68-a799ed33d02f, 'name': SearchDatastore_Task, 'duration_secs': 0.05114} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.845297] env[63202]: DEBUG oslo_concurrency.lockutils [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.845432] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb/0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 882.845647] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.845830] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 882.846050] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7a1b666c-7f85-4358-bc02-a0be21f180b8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.847801] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b4ea57f7-1df1-4d45-96fe-238f33e3a2a8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.853461] env[63202]: DEBUG oslo_vmware.api [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Waiting for the task: (returnval){ [ 882.853461] env[63202]: value = "task-1385388" [ 882.853461] env[63202]: _type = "Task" [ 882.853461] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.856930] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 882.857108] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 882.858048] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9801028d-7476-475e-ab8c-c16e3181ab02 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.862801] env[63202]: DEBUG oslo_vmware.api [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385388, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.865309] env[63202]: DEBUG oslo_vmware.api [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Waiting for the task: (returnval){ [ 882.865309] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52493f1c-dd87-2479-a37a-049fb96f78a9" [ 882.865309] env[63202]: _type = "Task" [ 882.865309] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.872513] env[63202]: DEBUG oslo_vmware.api [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52493f1c-dd87-2479-a37a-049fb96f78a9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.065039] env[63202]: DEBUG nova.scheduler.client.report [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 883.212477] env[63202]: DEBUG nova.compute.manager [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 883.237319] env[63202]: DEBUG nova.virt.hardware [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 883.237642] env[63202]: DEBUG nova.virt.hardware [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 883.237840] env[63202]: DEBUG nova.virt.hardware [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 883.238075] env[63202]: DEBUG nova.virt.hardware [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 883.238259] env[63202]: DEBUG nova.virt.hardware [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 883.238413] env[63202]: DEBUG nova.virt.hardware [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 883.238619] env[63202]: DEBUG nova.virt.hardware [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 883.238775] env[63202]: DEBUG nova.virt.hardware [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 883.239246] env[63202]: DEBUG nova.virt.hardware [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 883.239246] env[63202]: DEBUG nova.virt.hardware [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 883.239352] env[63202]: DEBUG nova.virt.hardware [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 883.240202] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5eb10eb-4a1f-412b-bd72-d4da7b3b0ef2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.248985] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7616283-069a-4e8e-8b07-d2fbd24d2786 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.253276] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 883.253504] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager.update_available_resource {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 883.291060] env[63202]: DEBUG oslo_vmware.api [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Task: {'id': task-1385387, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063144} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.291060] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 883.291887] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37aa1567-0b53-4767-9de2-f98a63426394 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.314979] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Reconfiguring VM instance instance-0000003e to attach disk [datastore1] cedc3a06-2123-4c5e-a6c3-599a3efc3c65/cedc3a06-2123-4c5e-a6c3-599a3efc3c65.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 883.315311] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b26997e1-6eb5-465f-a32d-d23a024790ef {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.335206] env[63202]: DEBUG oslo_vmware.api [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Waiting for the task: (returnval){ [ 883.335206] env[63202]: value = "task-1385389" [ 883.335206] env[63202]: _type = "Task" [ 883.335206] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.342856] env[63202]: DEBUG oslo_vmware.api [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Task: {'id': task-1385389, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.361833] env[63202]: DEBUG oslo_vmware.api [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385388, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.466497} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.362126] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb/0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 883.362352] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 883.362588] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-270402ae-56af-4656-8732-ae9d7085b900 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.369329] env[63202]: DEBUG oslo_vmware.api [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Waiting for the task: (returnval){ [ 883.369329] env[63202]: value = "task-1385390" [ 883.369329] env[63202]: _type = "Task" [ 883.369329] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.376466] env[63202]: DEBUG oslo_vmware.api [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52493f1c-dd87-2479-a37a-049fb96f78a9, 'name': SearchDatastore_Task, 'duration_secs': 0.014884} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.377827] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db5b70ee-c741-4b5e-b166-59100662017f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.382708] env[63202]: DEBUG oslo_vmware.api [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385390, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.385549] env[63202]: DEBUG oslo_vmware.api [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Waiting for the task: (returnval){ [ 883.385549] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]5235903e-7aea-2dd4-13d6-823813e36939" [ 883.385549] env[63202]: _type = "Task" [ 883.385549] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.392845] env[63202]: DEBUG oslo_vmware.api [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]5235903e-7aea-2dd4-13d6-823813e36939, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.569927] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.380s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.571944] env[63202]: DEBUG nova.compute.manager [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 883.574072] env[63202]: DEBUG oslo_concurrency.lockutils [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.248s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.575859] env[63202]: INFO nova.compute.claims [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 883.650897] env[63202]: DEBUG nova.compute.manager [req-41aeecff-117b-4715-bd08-822ae70e3cd9 req-d589da58-f3fb-4aa1-b942-312a51e67c63 service nova] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Received event network-vif-plugged-d12e501a-1681-4490-84f1-f48efcf89671 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 883.651135] env[63202]: DEBUG oslo_concurrency.lockutils [req-41aeecff-117b-4715-bd08-822ae70e3cd9 req-d589da58-f3fb-4aa1-b942-312a51e67c63 service nova] Acquiring lock "5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.651342] env[63202]: DEBUG oslo_concurrency.lockutils [req-41aeecff-117b-4715-bd08-822ae70e3cd9 req-d589da58-f3fb-4aa1-b942-312a51e67c63 service nova] Lock "5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.651506] env[63202]: DEBUG oslo_concurrency.lockutils [req-41aeecff-117b-4715-bd08-822ae70e3cd9 req-d589da58-f3fb-4aa1-b942-312a51e67c63 service nova] Lock "5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.651670] env[63202]: DEBUG nova.compute.manager [req-41aeecff-117b-4715-bd08-822ae70e3cd9 req-d589da58-f3fb-4aa1-b942-312a51e67c63 service nova] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] No waiting events found dispatching network-vif-plugged-d12e501a-1681-4490-84f1-f48efcf89671 {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 883.651833] env[63202]: WARNING nova.compute.manager [req-41aeecff-117b-4715-bd08-822ae70e3cd9 req-d589da58-f3fb-4aa1-b942-312a51e67c63 service nova] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Received unexpected event network-vif-plugged-d12e501a-1681-4490-84f1-f48efcf89671 for instance with vm_state building and task_state spawning. [ 883.703395] env[63202]: DEBUG nova.network.neutron [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Successfully updated port: d12e501a-1681-4490-84f1-f48efcf89671 {{(pid=63202) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 883.756997] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.846227] env[63202]: DEBUG oslo_vmware.api [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Task: {'id': task-1385389, 'name': ReconfigVM_Task, 'duration_secs': 0.298245} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.846569] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Reconfigured VM instance instance-0000003e to attach disk [datastore1] cedc3a06-2123-4c5e-a6c3-599a3efc3c65/cedc3a06-2123-4c5e-a6c3-599a3efc3c65.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 883.847161] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ffd12893-3d53-43fb-9a0b-78fe0018816f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.853620] env[63202]: DEBUG oslo_vmware.api [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Waiting for the task: (returnval){ [ 883.853620] env[63202]: value = "task-1385391" [ 883.853620] env[63202]: _type = "Task" [ 883.853620] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.861445] env[63202]: DEBUG oslo_vmware.api [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Task: {'id': task-1385391, 'name': Rename_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.878734] env[63202]: DEBUG oslo_vmware.api [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385390, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065782} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.878988] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 883.879792] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7b6908f-9ad2-40f0-883d-588300c292b8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.902660] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Reconfiguring VM instance instance-0000003f to attach disk [datastore1] 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb/0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 883.905692] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-27186074-a2a3-4daf-a9f1-534d73a3c8b1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.925278] env[63202]: DEBUG oslo_vmware.api [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]5235903e-7aea-2dd4-13d6-823813e36939, 'name': SearchDatastore_Task, 'duration_secs': 0.008971} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.925955] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.926226] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] a800fa5d-46a1-4a96-8ea2-26e34b869aa6/a800fa5d-46a1-4a96-8ea2-26e34b869aa6.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 883.926483] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-92c4e633-16da-4345-9f9a-ed6803f06157 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.929034] env[63202]: DEBUG oslo_vmware.api [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Waiting for the task: (returnval){ [ 883.929034] env[63202]: value = "task-1385392" [ 883.929034] env[63202]: _type = "Task" [ 883.929034] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.933366] env[63202]: DEBUG oslo_vmware.api [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Waiting for the task: (returnval){ [ 883.933366] env[63202]: value = "task-1385393" [ 883.933366] env[63202]: _type = "Task" [ 883.933366] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.939292] env[63202]: DEBUG oslo_vmware.api [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385392, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.943732] env[63202]: DEBUG oslo_vmware.api [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Task: {'id': task-1385393, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.081863] env[63202]: DEBUG nova.compute.utils [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 884.086412] env[63202]: DEBUG nova.compute.manager [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 884.086412] env[63202]: DEBUG nova.network.neutron [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 884.148651] env[63202]: DEBUG nova.policy [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2bd91973e1bb4ce7be8bce9bfe34a941', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e972faeaedc6468aab7e7cfee88a477b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 884.206716] env[63202]: DEBUG oslo_concurrency.lockutils [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Acquiring lock "refresh_cache-5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.206877] env[63202]: DEBUG oslo_concurrency.lockutils [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Acquired lock "refresh_cache-5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.207027] env[63202]: DEBUG nova.network.neutron [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 884.367304] env[63202]: DEBUG oslo_vmware.api [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Task: {'id': task-1385391, 'name': Rename_Task, 'duration_secs': 0.238299} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.369881] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 884.369881] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bf44c843-2793-4d34-81d4-1e702c8bec7a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.377021] env[63202]: DEBUG oslo_vmware.api [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Waiting for the task: (returnval){ [ 884.377021] env[63202]: value = "task-1385394" [ 884.377021] env[63202]: _type = "Task" [ 884.377021] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.382721] env[63202]: DEBUG oslo_vmware.api [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Task: {'id': task-1385394, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.453778] env[63202]: DEBUG oslo_vmware.api [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385392, 'name': ReconfigVM_Task, 'duration_secs': 0.289471} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.454484] env[63202]: DEBUG oslo_vmware.api [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Task: {'id': task-1385393, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.455155] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Reconfigured VM instance instance-0000003f to attach disk [datastore1] 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb/0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 884.456031] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a037fed9-d300-4aae-ac14-7420a4a8981f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.464124] env[63202]: DEBUG oslo_vmware.api [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Waiting for the task: (returnval){ [ 884.464124] env[63202]: value = "task-1385395" [ 884.464124] env[63202]: _type = "Task" [ 884.464124] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.473703] env[63202]: DEBUG oslo_vmware.api [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385395, 'name': Rename_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.489772] env[63202]: DEBUG nova.network.neutron [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Successfully created port: 6061b268-0c4c-414c-a0a3-a60196071be2 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 884.588665] env[63202]: DEBUG nova.compute.manager [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 884.762605] env[63202]: DEBUG nova.network.neutron [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 884.886873] env[63202]: DEBUG oslo_vmware.api [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Task: {'id': task-1385394, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.907934] env[63202]: DEBUG nova.network.neutron [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Updating instance_info_cache with network_info: [{"id": "d12e501a-1681-4490-84f1-f48efcf89671", "address": "fa:16:3e:90:d8:d2", "network": {"id": "bd667382-0ec4-4da0-9b78-f58a60972833", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.216", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "980e19d195ad46dbb597408ec0afcb9e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0ded8bac-871f-491b-94ec-cb67c08bc828", "external-id": "nsx-vlan-transportzone-212", "segmentation_id": 212, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd12e501a-16", "ovs_interfaceid": "d12e501a-1681-4490-84f1-f48efcf89671", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.935366] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64b07ee2-53f5-4060-ab70-86ffa5c2d64e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.949616] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-079541c7-b07b-42f9-ad3b-25e4dec44e5a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.953121] env[63202]: DEBUG oslo_vmware.api [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Task: {'id': task-1385393, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.821079} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.953434] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] a800fa5d-46a1-4a96-8ea2-26e34b869aa6/a800fa5d-46a1-4a96-8ea2-26e34b869aa6.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 884.953694] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] [instance: a800fa5d-46a1-4a96-8ea2-26e34b869aa6] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 884.954246] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-64ca1d1c-5901-4620-8eb3-77bb5d82d298 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.984492] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74436621-e6c4-4770-bab0-b1de5d4efb13 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.987168] env[63202]: DEBUG oslo_vmware.api [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Waiting for the task: (returnval){ [ 884.987168] env[63202]: value = "task-1385396" [ 884.987168] env[63202]: _type = "Task" [ 884.987168] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.997194] env[63202]: DEBUG oslo_vmware.api [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385395, 'name': Rename_Task, 'duration_secs': 0.149283} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.999443] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff1378be-5393-430e-8619-2b51889adc19 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.003526] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 885.005283] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-514294b3-e499-4aee-88ff-908b119c8771 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.011289] env[63202]: DEBUG oslo_vmware.api [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Task: {'id': task-1385396, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.020722] env[63202]: DEBUG nova.compute.provider_tree [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 885.022062] env[63202]: DEBUG oslo_vmware.api [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Waiting for the task: (returnval){ [ 885.022062] env[63202]: value = "task-1385397" [ 885.022062] env[63202]: _type = "Task" [ 885.022062] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.030471] env[63202]: DEBUG oslo_vmware.api [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385397, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.385971] env[63202]: DEBUG oslo_vmware.api [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Task: {'id': task-1385394, 'name': PowerOnVM_Task, 'duration_secs': 0.924831} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.386263] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 885.386460] env[63202]: INFO nova.compute.manager [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Took 9.24 seconds to spawn the instance on the hypervisor. [ 885.386633] env[63202]: DEBUG nova.compute.manager [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 885.387440] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9406aff-4ff6-4ddd-99ae-0d5695442830 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.410686] env[63202]: DEBUG oslo_concurrency.lockutils [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Releasing lock "refresh_cache-5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.411026] env[63202]: DEBUG nova.compute.manager [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Instance network_info: |[{"id": "d12e501a-1681-4490-84f1-f48efcf89671", "address": "fa:16:3e:90:d8:d2", "network": {"id": "bd667382-0ec4-4da0-9b78-f58a60972833", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.216", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "980e19d195ad46dbb597408ec0afcb9e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0ded8bac-871f-491b-94ec-cb67c08bc828", "external-id": "nsx-vlan-transportzone-212", "segmentation_id": 212, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd12e501a-16", "ovs_interfaceid": "d12e501a-1681-4490-84f1-f48efcf89671", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 885.411392] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:90:d8:d2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0ded8bac-871f-491b-94ec-cb67c08bc828', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd12e501a-1681-4490-84f1-f48efcf89671', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 885.419014] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Creating folder: Project (e2926b8bd92644deb614a386f00be6c8). Parent ref: group-v294090. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 885.419276] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9449ea37-877a-4fb2-a1f9-e5050452c7ff {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.429985] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Created folder: Project (e2926b8bd92644deb614a386f00be6c8) in parent group-v294090. [ 885.430186] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Creating folder: Instances. Parent ref: group-v294121. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 885.430415] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e338246a-1d9c-4944-a342-e3a5a437e030 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.440972] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Created folder: Instances in parent group-v294121. [ 885.441279] env[63202]: DEBUG oslo.service.loopingcall [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 885.441772] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 885.442041] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4e1e2dbf-c565-43be-b55e-7976868cb9e7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.464487] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 885.464487] env[63202]: value = "task-1385400" [ 885.464487] env[63202]: _type = "Task" [ 885.464487] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.472504] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385400, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.497442] env[63202]: DEBUG oslo_vmware.api [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Task: {'id': task-1385396, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06512} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.497893] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] [instance: a800fa5d-46a1-4a96-8ea2-26e34b869aa6] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 885.498712] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b01955c-b365-4914-bc64-29c919da48af {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.518985] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] [instance: a800fa5d-46a1-4a96-8ea2-26e34b869aa6] Reconfiguring VM instance instance-00000040 to attach disk [datastore1] a800fa5d-46a1-4a96-8ea2-26e34b869aa6/a800fa5d-46a1-4a96-8ea2-26e34b869aa6.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 885.519229] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0d87aa45-82be-45fa-a7f5-774e2af09cf0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.534323] env[63202]: DEBUG nova.scheduler.client.report [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 885.546505] env[63202]: DEBUG oslo_vmware.api [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385397, 'name': PowerOnVM_Task, 'duration_secs': 0.471133} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.547804] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 885.548027] env[63202]: INFO nova.compute.manager [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Took 7.02 seconds to spawn the instance on the hypervisor. [ 885.548377] env[63202]: DEBUG nova.compute.manager [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 885.548547] env[63202]: DEBUG oslo_vmware.api [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Waiting for the task: (returnval){ [ 885.548547] env[63202]: value = "task-1385401" [ 885.548547] env[63202]: _type = "Task" [ 885.548547] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.549832] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8681634b-f62a-43d8-9524-caf3a0bf2422 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.564837] env[63202]: DEBUG oslo_vmware.api [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Task: {'id': task-1385401, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.603451] env[63202]: DEBUG nova.compute.manager [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 885.630425] env[63202]: DEBUG nova.virt.hardware [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 885.630702] env[63202]: DEBUG nova.virt.hardware [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 885.630911] env[63202]: DEBUG nova.virt.hardware [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 885.631123] env[63202]: DEBUG nova.virt.hardware [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 885.631270] env[63202]: DEBUG nova.virt.hardware [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 885.631413] env[63202]: DEBUG nova.virt.hardware [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 885.631774] env[63202]: DEBUG nova.virt.hardware [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 885.631996] env[63202]: DEBUG nova.virt.hardware [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 885.632250] env[63202]: DEBUG nova.virt.hardware [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 885.632395] env[63202]: DEBUG nova.virt.hardware [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 885.632616] env[63202]: DEBUG nova.virt.hardware [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 885.633885] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1f106be-36ee-4d9e-a5ab-40e6fa00f178 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.642995] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a14ec97-2640-4b9b-892e-d8215c7122f8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.686512] env[63202]: DEBUG nova.compute.manager [req-10aec410-8f33-4d9e-b348-b63a5e4d324a req-0b2970ad-4f16-4306-ac0f-9650816cd395 service nova] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Received event network-changed-d12e501a-1681-4490-84f1-f48efcf89671 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 885.686730] env[63202]: DEBUG nova.compute.manager [req-10aec410-8f33-4d9e-b348-b63a5e4d324a req-0b2970ad-4f16-4306-ac0f-9650816cd395 service nova] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Refreshing instance network info cache due to event network-changed-d12e501a-1681-4490-84f1-f48efcf89671. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 885.686949] env[63202]: DEBUG oslo_concurrency.lockutils [req-10aec410-8f33-4d9e-b348-b63a5e4d324a req-0b2970ad-4f16-4306-ac0f-9650816cd395 service nova] Acquiring lock "refresh_cache-5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.687121] env[63202]: DEBUG oslo_concurrency.lockutils [req-10aec410-8f33-4d9e-b348-b63a5e4d324a req-0b2970ad-4f16-4306-ac0f-9650816cd395 service nova] Acquired lock "refresh_cache-5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.687296] env[63202]: DEBUG nova.network.neutron [req-10aec410-8f33-4d9e-b348-b63a5e4d324a req-0b2970ad-4f16-4306-ac0f-9650816cd395 service nova] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Refreshing network info cache for port d12e501a-1681-4490-84f1-f48efcf89671 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 885.910629] env[63202]: INFO nova.compute.manager [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Took 40.03 seconds to build instance. [ 885.975937] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385400, 'name': CreateVM_Task, 'duration_secs': 0.354699} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.976128] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 885.976838] env[63202]: DEBUG oslo_concurrency.lockutils [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.977015] env[63202]: DEBUG oslo_concurrency.lockutils [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.977336] env[63202]: DEBUG oslo_concurrency.lockutils [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 885.977589] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a20326b2-256e-4edf-a7ea-ec37362f4f09 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.982588] env[63202]: DEBUG oslo_vmware.api [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Waiting for the task: (returnval){ [ 885.982588] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52799bd5-5038-9c5d-7a3c-fd70712cfec6" [ 885.982588] env[63202]: _type = "Task" [ 885.982588] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.990702] env[63202]: DEBUG oslo_vmware.api [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52799bd5-5038-9c5d-7a3c-fd70712cfec6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.042825] env[63202]: DEBUG oslo_concurrency.lockutils [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.469s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.043355] env[63202]: DEBUG nova.compute.manager [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 886.046160] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.959s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.062289] env[63202]: DEBUG oslo_vmware.api [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Task: {'id': task-1385401, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.076439] env[63202]: INFO nova.compute.manager [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Took 38.45 seconds to build instance. [ 886.318216] env[63202]: DEBUG nova.network.neutron [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Successfully updated port: 6061b268-0c4c-414c-a0a3-a60196071be2 {{(pid=63202) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 886.412882] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b513d428-8767-417f-8540-5003c8a67acd tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Lock "cedc3a06-2123-4c5e-a6c3-599a3efc3c65" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 99.244s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.417696] env[63202]: DEBUG nova.network.neutron [req-10aec410-8f33-4d9e-b348-b63a5e4d324a req-0b2970ad-4f16-4306-ac0f-9650816cd395 service nova] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Updated VIF entry in instance network info cache for port d12e501a-1681-4490-84f1-f48efcf89671. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 886.417696] env[63202]: DEBUG nova.network.neutron [req-10aec410-8f33-4d9e-b348-b63a5e4d324a req-0b2970ad-4f16-4306-ac0f-9650816cd395 service nova] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Updating instance_info_cache with network_info: [{"id": "d12e501a-1681-4490-84f1-f48efcf89671", "address": "fa:16:3e:90:d8:d2", "network": {"id": "bd667382-0ec4-4da0-9b78-f58a60972833", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.216", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "980e19d195ad46dbb597408ec0afcb9e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0ded8bac-871f-491b-94ec-cb67c08bc828", "external-id": "nsx-vlan-transportzone-212", "segmentation_id": 212, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd12e501a-16", "ovs_interfaceid": "d12e501a-1681-4490-84f1-f48efcf89671", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.493191] env[63202]: DEBUG oslo_vmware.api [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52799bd5-5038-9c5d-7a3c-fd70712cfec6, 'name': SearchDatastore_Task, 'duration_secs': 0.091932} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.493499] env[63202]: DEBUG oslo_concurrency.lockutils [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.493752] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 886.494015] env[63202]: DEBUG oslo_concurrency.lockutils [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.494167] env[63202]: DEBUG oslo_concurrency.lockutils [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.494349] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 886.494914] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6ccbd110-7176-4687-8139-633d5f82f7fb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.503311] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 886.503479] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 886.504192] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a1c1eb1-1dd1-4cb4-bdce-9de52ac3234d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.509390] env[63202]: DEBUG oslo_vmware.api [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Waiting for the task: (returnval){ [ 886.509390] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52968b48-4f6e-a40d-d968-13334e8ec9bf" [ 886.509390] env[63202]: _type = "Task" [ 886.509390] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.516862] env[63202]: DEBUG oslo_vmware.api [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52968b48-4f6e-a40d-d968-13334e8ec9bf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.550919] env[63202]: DEBUG nova.compute.utils [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 886.552298] env[63202]: DEBUG nova.compute.manager [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 886.552432] env[63202]: DEBUG nova.network.neutron [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 886.566179] env[63202]: DEBUG oslo_vmware.api [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Task: {'id': task-1385401, 'name': ReconfigVM_Task, 'duration_secs': 0.930872} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.566179] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] [instance: a800fa5d-46a1-4a96-8ea2-26e34b869aa6] Reconfigured VM instance instance-00000040 to attach disk [datastore1] a800fa5d-46a1-4a96-8ea2-26e34b869aa6/a800fa5d-46a1-4a96-8ea2-26e34b869aa6.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 886.566648] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-39caf010-cbe2-4912-b824-b19dc0fc8dea {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.573426] env[63202]: DEBUG oslo_vmware.api [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Waiting for the task: (returnval){ [ 886.573426] env[63202]: value = "task-1385402" [ 886.573426] env[63202]: _type = "Task" [ 886.573426] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.579192] env[63202]: DEBUG oslo_concurrency.lockutils [None req-fe669d16-4a06-487b-8955-7d03bfc8363b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lock "0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 93.856s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.584507] env[63202]: DEBUG oslo_vmware.api [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Task: {'id': task-1385402, 'name': Rename_Task} progress is 6%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.612849] env[63202]: DEBUG nova.policy [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ca60dbf953fc42a5b6f95d06b8e3438d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e1a8d47e87d64f83b4b55d9704531021', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 886.824776] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "refresh_cache-50854a4e-8e4e-431b-8220-01bf8906b1c0" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.824776] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquired lock "refresh_cache-50854a4e-8e4e-431b-8220-01bf8906b1c0" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.824776] env[63202]: DEBUG nova.network.neutron [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 886.857923] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28aa138b-c31f-4a72-9461-65a33e3d9422 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.867180] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-065c6aad-6206-40a0-8027-d8b91343d4a0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.908759] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b50941c8-9376-49b6-88bc-2f48bb1e7498 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.916761] env[63202]: DEBUG nova.network.neutron [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Successfully created port: a0175615-5853-4224-8504-2a296446e39b {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 886.919765] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2973d3c-d266-46ae-9efd-1cfdd680893a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.923715] env[63202]: DEBUG nova.compute.manager [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 886.927039] env[63202]: DEBUG oslo_concurrency.lockutils [req-10aec410-8f33-4d9e-b348-b63a5e4d324a req-0b2970ad-4f16-4306-ac0f-9650816cd395 service nova] Releasing lock "refresh_cache-5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.941747] env[63202]: DEBUG nova.compute.provider_tree [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 887.019829] env[63202]: DEBUG oslo_vmware.api [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52968b48-4f6e-a40d-d968-13334e8ec9bf, 'name': SearchDatastore_Task, 'duration_secs': 0.012073} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.020767] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9e784ab1-a2de-4c45-850b-08d3a870fe99 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.026010] env[63202]: DEBUG oslo_vmware.api [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Waiting for the task: (returnval){ [ 887.026010] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52741f18-96e1-0e9f-982f-d2e8bf97c4a0" [ 887.026010] env[63202]: _type = "Task" [ 887.026010] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.034224] env[63202]: DEBUG oslo_vmware.api [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52741f18-96e1-0e9f-982f-d2e8bf97c4a0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.057500] env[63202]: DEBUG nova.compute.manager [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 887.083147] env[63202]: DEBUG nova.compute.manager [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 887.085671] env[63202]: DEBUG oslo_vmware.api [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Task: {'id': task-1385402, 'name': Rename_Task, 'duration_secs': 0.147232} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.087461] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] [instance: a800fa5d-46a1-4a96-8ea2-26e34b869aa6] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 887.087461] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4ae2c763-ae35-49e0-a3d5-f690db645e50 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.098300] env[63202]: DEBUG oslo_vmware.api [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Waiting for the task: (returnval){ [ 887.098300] env[63202]: value = "task-1385403" [ 887.098300] env[63202]: _type = "Task" [ 887.098300] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.106818] env[63202]: DEBUG oslo_vmware.api [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Task: {'id': task-1385403, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.389928] env[63202]: DEBUG nova.network.neutron [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 887.446889] env[63202]: DEBUG nova.scheduler.client.report [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 887.456704] env[63202]: DEBUG oslo_concurrency.lockutils [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.540283] env[63202]: DEBUG oslo_vmware.api [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52741f18-96e1-0e9f-982f-d2e8bf97c4a0, 'name': SearchDatastore_Task, 'duration_secs': 0.011148} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.540283] env[63202]: DEBUG oslo_concurrency.lockutils [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.540283] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a/5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 887.540283] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ef3299d0-a8a8-4284-8128-52e571f660d6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.546847] env[63202]: DEBUG oslo_vmware.api [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Waiting for the task: (returnval){ [ 887.546847] env[63202]: value = "task-1385404" [ 887.546847] env[63202]: _type = "Task" [ 887.546847] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.554583] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7a451f85-aa73-49f0-af5e-6452e182ee9b tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Acquiring lock "cedc3a06-2123-4c5e-a6c3-599a3efc3c65" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.554583] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7a451f85-aa73-49f0-af5e-6452e182ee9b tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Lock "cedc3a06-2123-4c5e-a6c3-599a3efc3c65" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.554583] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7a451f85-aa73-49f0-af5e-6452e182ee9b tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Acquiring lock "cedc3a06-2123-4c5e-a6c3-599a3efc3c65-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.554583] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7a451f85-aa73-49f0-af5e-6452e182ee9b tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Lock "cedc3a06-2123-4c5e-a6c3-599a3efc3c65-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.554713] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7a451f85-aa73-49f0-af5e-6452e182ee9b tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Lock "cedc3a06-2123-4c5e-a6c3-599a3efc3c65-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.561223] env[63202]: DEBUG oslo_vmware.api [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Task: {'id': task-1385404, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.561779] env[63202]: INFO nova.compute.manager [None req-7a451f85-aa73-49f0-af5e-6452e182ee9b tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Terminating instance [ 887.568904] env[63202]: DEBUG nova.compute.manager [None req-7a451f85-aa73-49f0-af5e-6452e182ee9b tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 887.568904] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-7a451f85-aa73-49f0-af5e-6452e182ee9b tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 887.568981] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41bc3511-3571-424b-98b0-6f908957c098 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.578167] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a451f85-aa73-49f0-af5e-6452e182ee9b tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 887.578423] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-83689336-6342-44c1-ab2f-9744a147fed7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.585809] env[63202]: DEBUG oslo_vmware.api [None req-7a451f85-aa73-49f0-af5e-6452e182ee9b tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Waiting for the task: (returnval){ [ 887.585809] env[63202]: value = "task-1385405" [ 887.585809] env[63202]: _type = "Task" [ 887.585809] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.601798] env[63202]: DEBUG oslo_vmware.api [None req-7a451f85-aa73-49f0-af5e-6452e182ee9b tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Task: {'id': task-1385405, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.613337] env[63202]: DEBUG oslo_vmware.api [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Task: {'id': task-1385403, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.614285] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.642949] env[63202]: DEBUG nova.network.neutron [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Updating instance_info_cache with network_info: [{"id": "6061b268-0c4c-414c-a0a3-a60196071be2", "address": "fa:16:3e:6c:4c:6c", "network": {"id": "b1ea2320-3801-43fe-b566-4f69db81329e", "bridge": "br-int", "label": "tempest-ServersTestJSON-618582355-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e972faeaedc6468aab7e7cfee88a477b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4223acd2-30f7-440e-b975-60b30d931694", "external-id": "nsx-vlan-transportzone-647", "segmentation_id": 647, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6061b268-0c", "ovs_interfaceid": "6061b268-0c4c-414c-a0a3-a60196071be2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.738550] env[63202]: DEBUG nova.compute.manager [req-acf44da5-44ee-4dee-94d4-b49daa57e09c req-86426e38-71cb-49dd-b7ed-d2d2f5be0e50 service nova] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Received event network-vif-plugged-6061b268-0c4c-414c-a0a3-a60196071be2 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 887.738831] env[63202]: DEBUG oslo_concurrency.lockutils [req-acf44da5-44ee-4dee-94d4-b49daa57e09c req-86426e38-71cb-49dd-b7ed-d2d2f5be0e50 service nova] Acquiring lock "50854a4e-8e4e-431b-8220-01bf8906b1c0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.739081] env[63202]: DEBUG oslo_concurrency.lockutils [req-acf44da5-44ee-4dee-94d4-b49daa57e09c req-86426e38-71cb-49dd-b7ed-d2d2f5be0e50 service nova] Lock "50854a4e-8e4e-431b-8220-01bf8906b1c0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.739451] env[63202]: DEBUG oslo_concurrency.lockutils [req-acf44da5-44ee-4dee-94d4-b49daa57e09c req-86426e38-71cb-49dd-b7ed-d2d2f5be0e50 service nova] Lock "50854a4e-8e4e-431b-8220-01bf8906b1c0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.739451] env[63202]: DEBUG nova.compute.manager [req-acf44da5-44ee-4dee-94d4-b49daa57e09c req-86426e38-71cb-49dd-b7ed-d2d2f5be0e50 service nova] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] No waiting events found dispatching network-vif-plugged-6061b268-0c4c-414c-a0a3-a60196071be2 {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 887.739580] env[63202]: WARNING nova.compute.manager [req-acf44da5-44ee-4dee-94d4-b49daa57e09c req-86426e38-71cb-49dd-b7ed-d2d2f5be0e50 service nova] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Received unexpected event network-vif-plugged-6061b268-0c4c-414c-a0a3-a60196071be2 for instance with vm_state building and task_state spawning. [ 887.739721] env[63202]: DEBUG nova.compute.manager [req-acf44da5-44ee-4dee-94d4-b49daa57e09c req-86426e38-71cb-49dd-b7ed-d2d2f5be0e50 service nova] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Received event network-changed-6061b268-0c4c-414c-a0a3-a60196071be2 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 887.740030] env[63202]: DEBUG nova.compute.manager [req-acf44da5-44ee-4dee-94d4-b49daa57e09c req-86426e38-71cb-49dd-b7ed-d2d2f5be0e50 service nova] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Refreshing instance network info cache due to event network-changed-6061b268-0c4c-414c-a0a3-a60196071be2. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 887.740193] env[63202]: DEBUG oslo_concurrency.lockutils [req-acf44da5-44ee-4dee-94d4-b49daa57e09c req-86426e38-71cb-49dd-b7ed-d2d2f5be0e50 service nova] Acquiring lock "refresh_cache-50854a4e-8e4e-431b-8220-01bf8906b1c0" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.952615] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.906s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.953466] env[63202]: ERROR nova.compute.manager [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e9f72a2a-5ce9-4802-bd15-9c22c86662b9, please check neutron logs for more information. [ 887.953466] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] Traceback (most recent call last): [ 887.953466] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 887.953466] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] self.driver.spawn(context, instance, image_meta, [ 887.953466] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 887.953466] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] self._vmops.spawn(context, instance, image_meta, injected_files, [ 887.953466] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 887.953466] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] vm_ref = self.build_virtual_machine(instance, [ 887.953466] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 887.953466] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] vif_infos = vmwarevif.get_vif_info(self._session, [ 887.953466] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 887.953796] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] for vif in network_info: [ 887.953796] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 887.953796] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] return self._sync_wrapper(fn, *args, **kwargs) [ 887.953796] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 887.953796] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] self.wait() [ 887.953796] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 887.953796] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] self[:] = self._gt.wait() [ 887.953796] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 887.953796] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] return self._exit_event.wait() [ 887.953796] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 887.953796] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] result = hub.switch() [ 887.953796] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 887.953796] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] return self.greenlet.switch() [ 887.954137] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 887.954137] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] result = function(*args, **kwargs) [ 887.954137] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 887.954137] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] return func(*args, **kwargs) [ 887.954137] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 887.954137] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] raise e [ 887.954137] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 887.954137] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] nwinfo = self.network_api.allocate_for_instance( [ 887.954137] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 887.954137] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] created_port_ids = self._update_ports_for_instance( [ 887.954137] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 887.954137] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] with excutils.save_and_reraise_exception(): [ 887.954137] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 887.954497] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] self.force_reraise() [ 887.954497] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 887.954497] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] raise self.value [ 887.954497] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 887.954497] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] updated_port = self._update_port( [ 887.954497] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 887.954497] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] _ensure_no_port_binding_failure(port) [ 887.954497] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 887.954497] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] raise exception.PortBindingFailed(port_id=port['id']) [ 887.954497] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] nova.exception.PortBindingFailed: Binding failed for port e9f72a2a-5ce9-4802-bd15-9c22c86662b9, please check neutron logs for more information. [ 887.954497] env[63202]: ERROR nova.compute.manager [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] [ 887.954802] env[63202]: DEBUG nova.compute.utils [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] Binding failed for port e9f72a2a-5ce9-4802-bd15-9c22c86662b9, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 887.955715] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.971s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.958838] env[63202]: DEBUG nova.compute.manager [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] Build of instance ac48e21a-72ec-4c9d-8262-5d6cb36ec531 was re-scheduled: Binding failed for port e9f72a2a-5ce9-4802-bd15-9c22c86662b9, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 887.959343] env[63202]: DEBUG nova.compute.manager [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 887.959579] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] Acquiring lock "refresh_cache-ac48e21a-72ec-4c9d-8262-5d6cb36ec531" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.959729] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] Acquired lock "refresh_cache-ac48e21a-72ec-4c9d-8262-5d6cb36ec531" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.959924] env[63202]: DEBUG nova.network.neutron [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 888.059116] env[63202]: DEBUG oslo_vmware.api [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Task: {'id': task-1385404, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.509639} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.059461] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a/5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 888.059718] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 888.060034] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bb67cd5e-aded-444a-93e9-633799c790c1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.068228] env[63202]: DEBUG oslo_vmware.api [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Waiting for the task: (returnval){ [ 888.068228] env[63202]: value = "task-1385406" [ 888.068228] env[63202]: _type = "Task" [ 888.068228] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.073119] env[63202]: DEBUG nova.compute.manager [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 888.081484] env[63202]: DEBUG oslo_vmware.api [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Task: {'id': task-1385406, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.098912] env[63202]: DEBUG oslo_vmware.api [None req-7a451f85-aa73-49f0-af5e-6452e182ee9b tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Task: {'id': task-1385405, 'name': PowerOffVM_Task, 'duration_secs': 0.201252} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.102085] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a451f85-aa73-49f0-af5e-6452e182ee9b tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 888.102266] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-7a451f85-aa73-49f0-af5e-6452e182ee9b tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 888.102516] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dfc3b9e7-7d48-479e-aeae-b5ff9ab99a52 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.105902] env[63202]: DEBUG nova.virt.hardware [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 888.106150] env[63202]: DEBUG nova.virt.hardware [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 888.106311] env[63202]: DEBUG nova.virt.hardware [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 888.106491] env[63202]: DEBUG nova.virt.hardware [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 888.106630] env[63202]: DEBUG nova.virt.hardware [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 888.106770] env[63202]: DEBUG nova.virt.hardware [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 888.106973] env[63202]: DEBUG nova.virt.hardware [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 888.107143] env[63202]: DEBUG nova.virt.hardware [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 888.107318] env[63202]: DEBUG nova.virt.hardware [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 888.107462] env[63202]: DEBUG nova.virt.hardware [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 888.107629] env[63202]: DEBUG nova.virt.hardware [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 888.108903] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3abdf020-9555-4ecc-b79f-b27f5cb32388 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.117389] env[63202]: DEBUG oslo_vmware.api [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Task: {'id': task-1385403, 'name': PowerOnVM_Task} progress is 78%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.122895] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aa47873-7fb4-40cd-b7e2-bc34a6925ab1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.144771] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Releasing lock "refresh_cache-50854a4e-8e4e-431b-8220-01bf8906b1c0" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.145172] env[63202]: DEBUG nova.compute.manager [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Instance network_info: |[{"id": "6061b268-0c4c-414c-a0a3-a60196071be2", "address": "fa:16:3e:6c:4c:6c", "network": {"id": "b1ea2320-3801-43fe-b566-4f69db81329e", "bridge": "br-int", "label": "tempest-ServersTestJSON-618582355-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e972faeaedc6468aab7e7cfee88a477b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4223acd2-30f7-440e-b975-60b30d931694", "external-id": "nsx-vlan-transportzone-647", "segmentation_id": 647, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6061b268-0c", "ovs_interfaceid": "6061b268-0c4c-414c-a0a3-a60196071be2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 888.145507] env[63202]: DEBUG oslo_concurrency.lockutils [req-acf44da5-44ee-4dee-94d4-b49daa57e09c req-86426e38-71cb-49dd-b7ed-d2d2f5be0e50 service nova] Acquired lock "refresh_cache-50854a4e-8e4e-431b-8220-01bf8906b1c0" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.145704] env[63202]: DEBUG nova.network.neutron [req-acf44da5-44ee-4dee-94d4-b49daa57e09c req-86426e38-71cb-49dd-b7ed-d2d2f5be0e50 service nova] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Refreshing network info cache for port 6061b268-0c4c-414c-a0a3-a60196071be2 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 888.147114] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6c:4c:6c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4223acd2-30f7-440e-b975-60b30d931694', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6061b268-0c4c-414c-a0a3-a60196071be2', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 888.154942] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Creating folder: Project (e972faeaedc6468aab7e7cfee88a477b). Parent ref: group-v294090. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 888.155532] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b925658a-4a11-4d5a-89f3-a0d325128145 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.167997] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Created folder: Project (e972faeaedc6468aab7e7cfee88a477b) in parent group-v294090. [ 888.168217] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Creating folder: Instances. Parent ref: group-v294124. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 888.168472] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2bbf4e8f-308b-4393-a2fc-16cb04d22d7a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.179317] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Created folder: Instances in parent group-v294124. [ 888.179868] env[63202]: DEBUG oslo.service.loopingcall [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 888.179868] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 888.180014] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ca301260-f83f-473e-a3ae-bfe675391ab1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.196414] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-7a451f85-aa73-49f0-af5e-6452e182ee9b tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 888.196658] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-7a451f85-aa73-49f0-af5e-6452e182ee9b tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Deleting contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 888.196879] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a451f85-aa73-49f0-af5e-6452e182ee9b tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Deleting the datastore file [datastore1] cedc3a06-2123-4c5e-a6c3-599a3efc3c65 {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 888.197577] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e5d5f81a-c95d-4aba-a4a6-87b8eb1ab722 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.202550] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 888.202550] env[63202]: value = "task-1385410" [ 888.202550] env[63202]: _type = "Task" [ 888.202550] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.208240] env[63202]: DEBUG oslo_vmware.api [None req-7a451f85-aa73-49f0-af5e-6452e182ee9b tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Waiting for the task: (returnval){ [ 888.208240] env[63202]: value = "task-1385411" [ 888.208240] env[63202]: _type = "Task" [ 888.208240] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.220582] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385410, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.223914] env[63202]: DEBUG oslo_vmware.api [None req-7a451f85-aa73-49f0-af5e-6452e182ee9b tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Task: {'id': task-1385411, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.487406] env[63202]: DEBUG nova.network.neutron [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 888.578738] env[63202]: DEBUG oslo_vmware.api [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Task: {'id': task-1385406, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082235} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.581134] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 888.582202] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-341770d9-833b-4d16-9ca6-9d90c6893e90 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.606824] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Reconfiguring VM instance instance-00000041 to attach disk [datastore1] 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a/5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 888.611125] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cfd35a43-6fb1-418e-a4db-e011bcd0cf8b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.627153] env[63202]: DEBUG nova.compute.manager [req-b38513a4-5a2f-4d09-937f-3935fd9b8827 req-97538de7-a4a6-4dd9-a4f9-a4a9e9010aa8 service nova] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Received event network-vif-plugged-a0175615-5853-4224-8504-2a296446e39b {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 888.627371] env[63202]: DEBUG oslo_concurrency.lockutils [req-b38513a4-5a2f-4d09-937f-3935fd9b8827 req-97538de7-a4a6-4dd9-a4f9-a4a9e9010aa8 service nova] Acquiring lock "2243bb69-0dc5-49cd-b94e-73e703cbadc0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.627573] env[63202]: DEBUG oslo_concurrency.lockutils [req-b38513a4-5a2f-4d09-937f-3935fd9b8827 req-97538de7-a4a6-4dd9-a4f9-a4a9e9010aa8 service nova] Lock "2243bb69-0dc5-49cd-b94e-73e703cbadc0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.627737] env[63202]: DEBUG oslo_concurrency.lockutils [req-b38513a4-5a2f-4d09-937f-3935fd9b8827 req-97538de7-a4a6-4dd9-a4f9-a4a9e9010aa8 service nova] Lock "2243bb69-0dc5-49cd-b94e-73e703cbadc0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.627900] env[63202]: DEBUG nova.compute.manager [req-b38513a4-5a2f-4d09-937f-3935fd9b8827 req-97538de7-a4a6-4dd9-a4f9-a4a9e9010aa8 service nova] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] No waiting events found dispatching network-vif-plugged-a0175615-5853-4224-8504-2a296446e39b {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 888.628075] env[63202]: WARNING nova.compute.manager [req-b38513a4-5a2f-4d09-937f-3935fd9b8827 req-97538de7-a4a6-4dd9-a4f9-a4a9e9010aa8 service nova] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Received unexpected event network-vif-plugged-a0175615-5853-4224-8504-2a296446e39b for instance with vm_state building and task_state spawning. [ 888.629820] env[63202]: DEBUG nova.network.neutron [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.656143] env[63202]: DEBUG oslo_vmware.api [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Waiting for the task: (returnval){ [ 888.656143] env[63202]: value = "task-1385412" [ 888.656143] env[63202]: _type = "Task" [ 888.656143] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.658582] env[63202]: DEBUG oslo_vmware.api [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Task: {'id': task-1385403, 'name': PowerOnVM_Task, 'duration_secs': 1.175198} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.659202] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] [instance: a800fa5d-46a1-4a96-8ea2-26e34b869aa6] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 888.659423] env[63202]: INFO nova.compute.manager [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] [instance: a800fa5d-46a1-4a96-8ea2-26e34b869aa6] Took 7.78 seconds to spawn the instance on the hypervisor. [ 888.659674] env[63202]: DEBUG nova.compute.manager [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] [instance: a800fa5d-46a1-4a96-8ea2-26e34b869aa6] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 888.664071] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4239b00-50d1-4d43-8658-2536b9e92c80 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.669445] env[63202]: DEBUG nova.network.neutron [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Successfully updated port: a0175615-5853-4224-8504-2a296446e39b {{(pid=63202) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 888.680470] env[63202]: DEBUG oslo_vmware.api [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Task: {'id': task-1385412, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.722370] env[63202]: DEBUG oslo_vmware.api [None req-7a451f85-aa73-49f0-af5e-6452e182ee9b tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Task: {'id': task-1385411, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.445939} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.728752] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a451f85-aa73-49f0-af5e-6452e182ee9b tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 888.729076] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-7a451f85-aa73-49f0-af5e-6452e182ee9b tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Deleted contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 888.729148] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-7a451f85-aa73-49f0-af5e-6452e182ee9b tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 888.729334] env[63202]: INFO nova.compute.manager [None req-7a451f85-aa73-49f0-af5e-6452e182ee9b tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Took 1.16 seconds to destroy the instance on the hypervisor. [ 888.729604] env[63202]: DEBUG oslo.service.loopingcall [None req-7a451f85-aa73-49f0-af5e-6452e182ee9b tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 888.730151] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385410, 'name': CreateVM_Task} progress is 25%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.732927] env[63202]: DEBUG nova.compute.manager [-] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 888.733068] env[63202]: DEBUG nova.network.neutron [-] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 888.837228] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93d9745f-238f-4bc0-935f-d48aabde0fd4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.845458] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b038762c-8e36-4be2-afc8-290a3a4f6103 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.878214] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1969f5da-6237-4cd8-8fa4-729ddc3c8f7b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.886853] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5071c541-62d0-4b8d-9e77-bb7ddbcebdf6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.902299] env[63202]: DEBUG nova.compute.provider_tree [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 889.142775] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] Releasing lock "refresh_cache-ac48e21a-72ec-4c9d-8262-5d6cb36ec531" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.143112] env[63202]: DEBUG nova.compute.manager [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 889.143227] env[63202]: DEBUG nova.compute.manager [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 889.143395] env[63202]: DEBUG nova.network.neutron [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 889.158705] env[63202]: DEBUG nova.network.neutron [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 889.170945] env[63202]: DEBUG oslo_vmware.api [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Task: {'id': task-1385412, 'name': ReconfigVM_Task, 'duration_secs': 0.305957} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.171932] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Reconfigured VM instance instance-00000041 to attach disk [datastore1] 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a/5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 889.172821] env[63202]: DEBUG oslo_concurrency.lockutils [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Acquiring lock "refresh_cache-2243bb69-0dc5-49cd-b94e-73e703cbadc0" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.172949] env[63202]: DEBUG oslo_concurrency.lockutils [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Acquired lock "refresh_cache-2243bb69-0dc5-49cd-b94e-73e703cbadc0" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.177097] env[63202]: DEBUG nova.network.neutron [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 889.178112] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9e0fe8db-7d9c-4107-bad7-56b7198f52c9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.186993] env[63202]: DEBUG oslo_vmware.api [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Waiting for the task: (returnval){ [ 889.186993] env[63202]: value = "task-1385413" [ 889.186993] env[63202]: _type = "Task" [ 889.186993] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.201138] env[63202]: INFO nova.compute.manager [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] [instance: a800fa5d-46a1-4a96-8ea2-26e34b869aa6] Took 39.76 seconds to build instance. [ 889.203066] env[63202]: DEBUG nova.network.neutron [req-acf44da5-44ee-4dee-94d4-b49daa57e09c req-86426e38-71cb-49dd-b7ed-d2d2f5be0e50 service nova] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Updated VIF entry in instance network info cache for port 6061b268-0c4c-414c-a0a3-a60196071be2. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 889.203371] env[63202]: DEBUG nova.network.neutron [req-acf44da5-44ee-4dee-94d4-b49daa57e09c req-86426e38-71cb-49dd-b7ed-d2d2f5be0e50 service nova] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Updating instance_info_cache with network_info: [{"id": "6061b268-0c4c-414c-a0a3-a60196071be2", "address": "fa:16:3e:6c:4c:6c", "network": {"id": "b1ea2320-3801-43fe-b566-4f69db81329e", "bridge": "br-int", "label": "tempest-ServersTestJSON-618582355-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e972faeaedc6468aab7e7cfee88a477b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4223acd2-30f7-440e-b975-60b30d931694", "external-id": "nsx-vlan-transportzone-647", "segmentation_id": 647, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6061b268-0c", "ovs_interfaceid": "6061b268-0c4c-414c-a0a3-a60196071be2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.214148] env[63202]: DEBUG oslo_vmware.api [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Task: {'id': task-1385413, 'name': Rename_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.218505] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385410, 'name': CreateVM_Task} progress is 99%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.405963] env[63202]: DEBUG nova.scheduler.client.report [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 889.495026] env[63202]: DEBUG nova.network.neutron [-] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.665956] env[63202]: DEBUG nova.network.neutron [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.697952] env[63202]: DEBUG oslo_vmware.api [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Task: {'id': task-1385413, 'name': Rename_Task, 'duration_secs': 0.151442} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.698487] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 889.698487] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-63ae2363-5a00-4c64-920f-4f3e534b54ed {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.702876] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4d50ef86-7652-4fb1-9c46-585afc595c81 tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Lock "a800fa5d-46a1-4a96-8ea2-26e34b869aa6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 96.943s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.705582] env[63202]: DEBUG oslo_concurrency.lockutils [req-acf44da5-44ee-4dee-94d4-b49daa57e09c req-86426e38-71cb-49dd-b7ed-d2d2f5be0e50 service nova] Releasing lock "refresh_cache-50854a4e-8e4e-431b-8220-01bf8906b1c0" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.705810] env[63202]: DEBUG nova.compute.manager [req-acf44da5-44ee-4dee-94d4-b49daa57e09c req-86426e38-71cb-49dd-b7ed-d2d2f5be0e50 service nova] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Received event network-changed-e5621970-d0a8-4803-94ef-f6d577bdf6d2 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 889.705977] env[63202]: DEBUG nova.compute.manager [req-acf44da5-44ee-4dee-94d4-b49daa57e09c req-86426e38-71cb-49dd-b7ed-d2d2f5be0e50 service nova] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Refreshing instance network info cache due to event network-changed-e5621970-d0a8-4803-94ef-f6d577bdf6d2. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 889.706254] env[63202]: DEBUG oslo_concurrency.lockutils [req-acf44da5-44ee-4dee-94d4-b49daa57e09c req-86426e38-71cb-49dd-b7ed-d2d2f5be0e50 service nova] Acquiring lock "refresh_cache-0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.706370] env[63202]: DEBUG oslo_concurrency.lockutils [req-acf44da5-44ee-4dee-94d4-b49daa57e09c req-86426e38-71cb-49dd-b7ed-d2d2f5be0e50 service nova] Acquired lock "refresh_cache-0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.706529] env[63202]: DEBUG nova.network.neutron [req-acf44da5-44ee-4dee-94d4-b49daa57e09c req-86426e38-71cb-49dd-b7ed-d2d2f5be0e50 service nova] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Refreshing network info cache for port e5621970-d0a8-4803-94ef-f6d577bdf6d2 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 889.709019] env[63202]: DEBUG oslo_vmware.api [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Waiting for the task: (returnval){ [ 889.709019] env[63202]: value = "task-1385414" [ 889.709019] env[63202]: _type = "Task" [ 889.709019] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.724253] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385410, 'name': CreateVM_Task} progress is 99%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.727545] env[63202]: DEBUG oslo_vmware.api [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Task: {'id': task-1385414, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.742704] env[63202]: DEBUG nova.network.neutron [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 889.911434] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.956s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.912173] env[63202]: ERROR nova.compute.manager [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6b7258a2-07d2-4cb5-b458-04ed148d1895, please check neutron logs for more information. [ 889.912173] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] Traceback (most recent call last): [ 889.912173] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 889.912173] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] self.driver.spawn(context, instance, image_meta, [ 889.912173] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 889.912173] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] self._vmops.spawn(context, instance, image_meta, injected_files, [ 889.912173] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 889.912173] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] vm_ref = self.build_virtual_machine(instance, [ 889.912173] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 889.912173] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] vif_infos = vmwarevif.get_vif_info(self._session, [ 889.912173] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 889.912497] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] for vif in network_info: [ 889.912497] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 889.912497] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] return self._sync_wrapper(fn, *args, **kwargs) [ 889.912497] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 889.912497] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] self.wait() [ 889.912497] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 889.912497] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] self[:] = self._gt.wait() [ 889.912497] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 889.912497] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] return self._exit_event.wait() [ 889.912497] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 889.912497] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] current.throw(*self._exc) [ 889.912497] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 889.912497] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] result = function(*args, **kwargs) [ 889.912839] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 889.912839] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] return func(*args, **kwargs) [ 889.912839] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 889.912839] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] raise e [ 889.912839] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 889.912839] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] nwinfo = self.network_api.allocate_for_instance( [ 889.912839] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 889.912839] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] created_port_ids = self._update_ports_for_instance( [ 889.912839] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 889.912839] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] with excutils.save_and_reraise_exception(): [ 889.912839] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 889.912839] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] self.force_reraise() [ 889.912839] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 889.913161] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] raise self.value [ 889.913161] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 889.913161] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] updated_port = self._update_port( [ 889.913161] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 889.913161] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] _ensure_no_port_binding_failure(port) [ 889.913161] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 889.913161] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] raise exception.PortBindingFailed(port_id=port['id']) [ 889.913161] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] nova.exception.PortBindingFailed: Binding failed for port 6b7258a2-07d2-4cb5-b458-04ed148d1895, please check neutron logs for more information. [ 889.913161] env[63202]: ERROR nova.compute.manager [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] [ 889.913161] env[63202]: DEBUG nova.compute.utils [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] Binding failed for port 6b7258a2-07d2-4cb5-b458-04ed148d1895, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 889.915316] env[63202]: DEBUG nova.network.neutron [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Updating instance_info_cache with network_info: [{"id": "a0175615-5853-4224-8504-2a296446e39b", "address": "fa:16:3e:d8:ab:76", "network": {"id": "334c5b4e-9931-457d-b98c-1df0cf4ccfa9", "bridge": "br-int", "label": "tempest-ServersTestJSON-720696570-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1a8d47e87d64f83b4b55d9704531021", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "399f3826-705c-45f7-9fe0-3a08a945151a", "external-id": "nsx-vlan-transportzone-936", "segmentation_id": 936, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0175615-58", "ovs_interfaceid": "a0175615-5853-4224-8504-2a296446e39b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.917200] env[63202]: DEBUG oslo_concurrency.lockutils [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.304s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.920283] env[63202]: DEBUG nova.compute.manager [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] Build of instance 69c66a35-3789-4c2f-b7cf-5f4136d53b94 was re-scheduled: Binding failed for port 6b7258a2-07d2-4cb5-b458-04ed148d1895, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 889.920780] env[63202]: DEBUG nova.compute.manager [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 889.921053] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] Acquiring lock "refresh_cache-69c66a35-3789-4c2f-b7cf-5f4136d53b94" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.921241] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] Acquired lock "refresh_cache-69c66a35-3789-4c2f-b7cf-5f4136d53b94" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.921439] env[63202]: DEBUG nova.network.neutron [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 889.997890] env[63202]: INFO nova.compute.manager [-] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Took 1.26 seconds to deallocate network for instance. [ 890.050051] env[63202]: DEBUG nova.compute.manager [None req-1f54c5aa-21c2-4af5-9f98-bdb4568e182d tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] [instance: a800fa5d-46a1-4a96-8ea2-26e34b869aa6] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 890.051176] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2613a144-2c20-4e19-99df-9f5582663278 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.168343] env[63202]: INFO nova.compute.manager [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] [instance: ac48e21a-72ec-4c9d-8262-5d6cb36ec531] Took 1.02 seconds to deallocate network for instance. [ 890.175282] env[63202]: DEBUG oslo_concurrency.lockutils [None req-0915f74e-cfc6-45f1-9f44-9af3c69affef tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Acquiring lock "a800fa5d-46a1-4a96-8ea2-26e34b869aa6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.175398] env[63202]: DEBUG oslo_concurrency.lockutils [None req-0915f74e-cfc6-45f1-9f44-9af3c69affef tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Lock "a800fa5d-46a1-4a96-8ea2-26e34b869aa6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.175598] env[63202]: DEBUG oslo_concurrency.lockutils [None req-0915f74e-cfc6-45f1-9f44-9af3c69affef tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Acquiring lock "a800fa5d-46a1-4a96-8ea2-26e34b869aa6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.175771] env[63202]: DEBUG oslo_concurrency.lockutils [None req-0915f74e-cfc6-45f1-9f44-9af3c69affef tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Lock "a800fa5d-46a1-4a96-8ea2-26e34b869aa6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.175933] env[63202]: DEBUG oslo_concurrency.lockutils [None req-0915f74e-cfc6-45f1-9f44-9af3c69affef tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Lock "a800fa5d-46a1-4a96-8ea2-26e34b869aa6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.177821] env[63202]: INFO nova.compute.manager [None req-0915f74e-cfc6-45f1-9f44-9af3c69affef tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] [instance: a800fa5d-46a1-4a96-8ea2-26e34b869aa6] Terminating instance [ 890.179377] env[63202]: DEBUG oslo_concurrency.lockutils [None req-0915f74e-cfc6-45f1-9f44-9af3c69affef tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Acquiring lock "refresh_cache-a800fa5d-46a1-4a96-8ea2-26e34b869aa6" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.179532] env[63202]: DEBUG oslo_concurrency.lockutils [None req-0915f74e-cfc6-45f1-9f44-9af3c69affef tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Acquired lock "refresh_cache-a800fa5d-46a1-4a96-8ea2-26e34b869aa6" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.179704] env[63202]: DEBUG nova.network.neutron [None req-0915f74e-cfc6-45f1-9f44-9af3c69affef tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] [instance: a800fa5d-46a1-4a96-8ea2-26e34b869aa6] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 890.206194] env[63202]: DEBUG nova.compute.manager [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 890.221116] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385410, 'name': CreateVM_Task, 'duration_secs': 1.713002} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.222031] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 890.222984] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.223372] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.223690] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 890.226822] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-714344d6-50b8-46db-b57e-87adf5a61d7b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.228395] env[63202]: DEBUG oslo_vmware.api [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Task: {'id': task-1385414, 'name': PowerOnVM_Task, 'duration_secs': 0.455425} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.228627] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 890.228811] env[63202]: INFO nova.compute.manager [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Took 7.02 seconds to spawn the instance on the hypervisor. [ 890.228983] env[63202]: DEBUG nova.compute.manager [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 890.230137] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dbd8271-44fd-4088-9ded-80f7672b4c25 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.233509] env[63202]: DEBUG oslo_vmware.api [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 890.233509] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52ca41bc-bd76-0fe6-74d7-1ccf7f40adf3" [ 890.233509] env[63202]: _type = "Task" [ 890.233509] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.249293] env[63202]: DEBUG oslo_vmware.api [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52ca41bc-bd76-0fe6-74d7-1ccf7f40adf3, 'name': SearchDatastore_Task, 'duration_secs': 0.012765} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.249572] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.249803] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 890.250047] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.250196] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.250371] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 890.250614] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a94f5973-ca8c-4c03-b8cc-4152aa1ea50a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.260581] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 890.260759] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 890.261897] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0b3a985a-f382-4461-87e0-162f652e0137 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.267432] env[63202]: DEBUG oslo_vmware.api [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 890.267432] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]5213787c-613d-d10f-365f-505904951bf4" [ 890.267432] env[63202]: _type = "Task" [ 890.267432] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.275941] env[63202]: DEBUG oslo_vmware.api [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]5213787c-613d-d10f-365f-505904951bf4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.421312] env[63202]: DEBUG oslo_concurrency.lockutils [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Releasing lock "refresh_cache-2243bb69-0dc5-49cd-b94e-73e703cbadc0" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.421527] env[63202]: DEBUG nova.compute.manager [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Instance network_info: |[{"id": "a0175615-5853-4224-8504-2a296446e39b", "address": "fa:16:3e:d8:ab:76", "network": {"id": "334c5b4e-9931-457d-b98c-1df0cf4ccfa9", "bridge": "br-int", "label": "tempest-ServersTestJSON-720696570-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1a8d47e87d64f83b4b55d9704531021", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "399f3826-705c-45f7-9fe0-3a08a945151a", "external-id": "nsx-vlan-transportzone-936", "segmentation_id": 936, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0175615-58", "ovs_interfaceid": "a0175615-5853-4224-8504-2a296446e39b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 890.421934] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d8:ab:76', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '399f3826-705c-45f7-9fe0-3a08a945151a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a0175615-5853-4224-8504-2a296446e39b', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 890.430200] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Creating folder: Project (e1a8d47e87d64f83b4b55d9704531021). Parent ref: group-v294090. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 890.431098] env[63202]: DEBUG nova.network.neutron [req-acf44da5-44ee-4dee-94d4-b49daa57e09c req-86426e38-71cb-49dd-b7ed-d2d2f5be0e50 service nova] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Updated VIF entry in instance network info cache for port e5621970-d0a8-4803-94ef-f6d577bdf6d2. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 890.431423] env[63202]: DEBUG nova.network.neutron [req-acf44da5-44ee-4dee-94d4-b49daa57e09c req-86426e38-71cb-49dd-b7ed-d2d2f5be0e50 service nova] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Updating instance_info_cache with network_info: [{"id": "e5621970-d0a8-4803-94ef-f6d577bdf6d2", "address": "fa:16:3e:00:ba:72", "network": {"id": "345538f4-e791-4fc6-9719-237c11f1382a", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-449651008-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.170", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a9e939cdf4a94a8aa0107a5761771c2c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa09e855-8af1-419b-b78d-8ffcc94b1bfb", "external-id": "nsx-vlan-transportzone-901", "segmentation_id": 901, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape5621970-d0", "ovs_interfaceid": "e5621970-d0a8-4803-94ef-f6d577bdf6d2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.437261] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f1683c6f-d315-4f19-8699-271876ffdebc {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.451411] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Created folder: Project (e1a8d47e87d64f83b4b55d9704531021) in parent group-v294090. [ 890.451968] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Creating folder: Instances. Parent ref: group-v294127. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 890.451968] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ba191588-9d8f-452b-a955-52bfd9cc2757 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.454824] env[63202]: DEBUG nova.network.neutron [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 890.465210] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Created folder: Instances in parent group-v294127. [ 890.466925] env[63202]: DEBUG oslo.service.loopingcall [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 890.466925] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 890.466925] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c701022b-6201-4d4c-8cbe-962c8077e23e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.487363] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 890.487363] env[63202]: value = "task-1385417" [ 890.487363] env[63202]: _type = "Task" [ 890.487363] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.495697] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385417, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.504917] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7a451f85-aa73-49f0-af5e-6452e182ee9b tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.532409] env[63202]: DEBUG nova.network.neutron [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.562767] env[63202]: INFO nova.compute.manager [None req-1f54c5aa-21c2-4af5-9f98-bdb4568e182d tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] [instance: a800fa5d-46a1-4a96-8ea2-26e34b869aa6] instance snapshotting [ 890.563416] env[63202]: DEBUG nova.objects.instance [None req-1f54c5aa-21c2-4af5-9f98-bdb4568e182d tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Lazy-loading 'flavor' on Instance uuid a800fa5d-46a1-4a96-8ea2-26e34b869aa6 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 890.623129] env[63202]: DEBUG nova.compute.manager [req-fdd4414f-e7f0-482f-bc71-2aa538d7db0a req-b6a213a3-6de2-4c1d-8deb-08889df576ae service nova] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Received event network-changed-a0175615-5853-4224-8504-2a296446e39b {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 890.623129] env[63202]: DEBUG nova.compute.manager [req-fdd4414f-e7f0-482f-bc71-2aa538d7db0a req-b6a213a3-6de2-4c1d-8deb-08889df576ae service nova] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Refreshing instance network info cache due to event network-changed-a0175615-5853-4224-8504-2a296446e39b. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 890.623129] env[63202]: DEBUG oslo_concurrency.lockutils [req-fdd4414f-e7f0-482f-bc71-2aa538d7db0a req-b6a213a3-6de2-4c1d-8deb-08889df576ae service nova] Acquiring lock "refresh_cache-2243bb69-0dc5-49cd-b94e-73e703cbadc0" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.623485] env[63202]: DEBUG oslo_concurrency.lockutils [req-fdd4414f-e7f0-482f-bc71-2aa538d7db0a req-b6a213a3-6de2-4c1d-8deb-08889df576ae service nova] Acquired lock "refresh_cache-2243bb69-0dc5-49cd-b94e-73e703cbadc0" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.623595] env[63202]: DEBUG nova.network.neutron [req-fdd4414f-e7f0-482f-bc71-2aa538d7db0a req-b6a213a3-6de2-4c1d-8deb-08889df576ae service nova] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Refreshing network info cache for port a0175615-5853-4224-8504-2a296446e39b {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 890.717197] env[63202]: DEBUG nova.network.neutron [None req-0915f74e-cfc6-45f1-9f44-9af3c69affef tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] [instance: a800fa5d-46a1-4a96-8ea2-26e34b869aa6] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 890.724584] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b939a17-82a5-4b82-abf9-2725d470ee8f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.731079] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.734080] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03b7968a-dbad-49d5-b598-480f3151ef40 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.781209] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc916a19-b553-44e1-b673-60b989e04b4b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.784972] env[63202]: DEBUG nova.network.neutron [None req-0915f74e-cfc6-45f1-9f44-9af3c69affef tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] [instance: a800fa5d-46a1-4a96-8ea2-26e34b869aa6] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.786359] env[63202]: INFO nova.compute.manager [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Took 39.45 seconds to build instance. [ 890.797785] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb4b1b63-5bec-4999-84bb-d290c5388f35 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.802298] env[63202]: DEBUG oslo_vmware.api [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]5213787c-613d-d10f-365f-505904951bf4, 'name': SearchDatastore_Task, 'duration_secs': 0.012816} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.804243] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c18b6326-5908-4b2d-a9fa-41c514903d1d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.816152] env[63202]: DEBUG nova.compute.provider_tree [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 890.821260] env[63202]: DEBUG oslo_vmware.api [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 890.821260] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]5221f321-efff-f5f5-3466-acb072d91100" [ 890.821260] env[63202]: _type = "Task" [ 890.821260] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.831420] env[63202]: DEBUG oslo_vmware.api [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]5221f321-efff-f5f5-3466-acb072d91100, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.940505] env[63202]: DEBUG oslo_concurrency.lockutils [req-acf44da5-44ee-4dee-94d4-b49daa57e09c req-86426e38-71cb-49dd-b7ed-d2d2f5be0e50 service nova] Releasing lock "refresh_cache-0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.999285] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385417, 'name': CreateVM_Task, 'duration_secs': 0.377631} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.999378] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 891.000132] env[63202]: DEBUG oslo_concurrency.lockutils [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.000295] env[63202]: DEBUG oslo_concurrency.lockutils [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.000655] env[63202]: DEBUG oslo_concurrency.lockutils [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 891.000930] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1bc9e069-adaa-44d8-928a-8a7e483f04fb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.006888] env[63202]: DEBUG oslo_vmware.api [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Waiting for the task: (returnval){ [ 891.006888] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52289216-ac17-b045-ba8a-ed22e47d10e3" [ 891.006888] env[63202]: _type = "Task" [ 891.006888] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.016932] env[63202]: DEBUG oslo_vmware.api [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52289216-ac17-b045-ba8a-ed22e47d10e3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.035650] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] Releasing lock "refresh_cache-69c66a35-3789-4c2f-b7cf-5f4136d53b94" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.035883] env[63202]: DEBUG nova.compute.manager [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 891.036077] env[63202]: DEBUG nova.compute.manager [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 891.036253] env[63202]: DEBUG nova.network.neutron [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 891.064152] env[63202]: DEBUG nova.network.neutron [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 891.069476] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee52cc1e-80ef-4bbb-82b6-47a9c1e3d658 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.088624] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5e31773-3b87-4839-a817-dc3833637ed4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.205639] env[63202]: INFO nova.scheduler.client.report [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] Deleted allocations for instance ac48e21a-72ec-4c9d-8262-5d6cb36ec531 [ 891.288547] env[63202]: DEBUG oslo_concurrency.lockutils [None req-0915f74e-cfc6-45f1-9f44-9af3c69affef tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Releasing lock "refresh_cache-a800fa5d-46a1-4a96-8ea2-26e34b869aa6" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.289023] env[63202]: DEBUG nova.compute.manager [None req-0915f74e-cfc6-45f1-9f44-9af3c69affef tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] [instance: a800fa5d-46a1-4a96-8ea2-26e34b869aa6] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 891.289238] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-0915f74e-cfc6-45f1-9f44-9af3c69affef tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] [instance: a800fa5d-46a1-4a96-8ea2-26e34b869aa6] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 891.290284] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46940f54-da40-4fd9-a491-000b2f10f03f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.293585] env[63202]: DEBUG oslo_concurrency.lockutils [None req-72d498ab-cad9-4d98-b9f2-fa8e235aac7e tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Lock "5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 97.496s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.300093] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-0915f74e-cfc6-45f1-9f44-9af3c69affef tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] [instance: a800fa5d-46a1-4a96-8ea2-26e34b869aa6] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 891.300276] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ce9d32c6-8e74-4228-a70a-3a13e0533cca {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.308634] env[63202]: DEBUG oslo_vmware.api [None req-0915f74e-cfc6-45f1-9f44-9af3c69affef tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Waiting for the task: (returnval){ [ 891.308634] env[63202]: value = "task-1385418" [ 891.308634] env[63202]: _type = "Task" [ 891.308634] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.322629] env[63202]: DEBUG nova.scheduler.client.report [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 891.328024] env[63202]: DEBUG oslo_vmware.api [None req-0915f74e-cfc6-45f1-9f44-9af3c69affef tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Task: {'id': task-1385418, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.339762] env[63202]: DEBUG oslo_vmware.api [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]5221f321-efff-f5f5-3466-acb072d91100, 'name': SearchDatastore_Task, 'duration_secs': 0.011209} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.340061] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.340672] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore2] 50854a4e-8e4e-431b-8220-01bf8906b1c0/50854a4e-8e4e-431b-8220-01bf8906b1c0.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 891.340960] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9853b4d4-c4d4-4034-a05d-fa6e4ac59af4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.350268] env[63202]: DEBUG oslo_vmware.api [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 891.350268] env[63202]: value = "task-1385419" [ 891.350268] env[63202]: _type = "Task" [ 891.350268] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.358462] env[63202]: DEBUG oslo_vmware.api [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385419, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.506408] env[63202]: DEBUG nova.network.neutron [req-fdd4414f-e7f0-482f-bc71-2aa538d7db0a req-b6a213a3-6de2-4c1d-8deb-08889df576ae service nova] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Updated VIF entry in instance network info cache for port a0175615-5853-4224-8504-2a296446e39b. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 891.506761] env[63202]: DEBUG nova.network.neutron [req-fdd4414f-e7f0-482f-bc71-2aa538d7db0a req-b6a213a3-6de2-4c1d-8deb-08889df576ae service nova] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Updating instance_info_cache with network_info: [{"id": "a0175615-5853-4224-8504-2a296446e39b", "address": "fa:16:3e:d8:ab:76", "network": {"id": "334c5b4e-9931-457d-b98c-1df0cf4ccfa9", "bridge": "br-int", "label": "tempest-ServersTestJSON-720696570-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1a8d47e87d64f83b4b55d9704531021", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "399f3826-705c-45f7-9fe0-3a08a945151a", "external-id": "nsx-vlan-transportzone-936", "segmentation_id": 936, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0175615-58", "ovs_interfaceid": "a0175615-5853-4224-8504-2a296446e39b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.519529] env[63202]: DEBUG oslo_vmware.api [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52289216-ac17-b045-ba8a-ed22e47d10e3, 'name': SearchDatastore_Task, 'duration_secs': 0.012986} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.520438] env[63202]: DEBUG oslo_concurrency.lockutils [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.520675] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 891.520996] env[63202]: DEBUG oslo_concurrency.lockutils [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.521163] env[63202]: DEBUG oslo_concurrency.lockutils [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.521356] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 891.521879] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2669bc0b-f600-49ae-af13-248d99dfedc5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.547507] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 891.547702] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 891.548542] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b949c92d-bab4-4fcc-a200-f9e7129d7384 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.558612] env[63202]: DEBUG oslo_vmware.api [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Waiting for the task: (returnval){ [ 891.558612] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]527bd600-7ee3-2316-b548-caacf5423d07" [ 891.558612] env[63202]: _type = "Task" [ 891.558612] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.567573] env[63202]: DEBUG oslo_vmware.api [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]527bd600-7ee3-2316-b548-caacf5423d07, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.569122] env[63202]: DEBUG nova.network.neutron [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.602703] env[63202]: DEBUG nova.compute.manager [None req-1f54c5aa-21c2-4af5-9f98-bdb4568e182d tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] [instance: a800fa5d-46a1-4a96-8ea2-26e34b869aa6] Instance disappeared during snapshot {{(pid=63202) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 891.686287] env[63202]: DEBUG oslo_concurrency.lockutils [None req-50e06c6a-8e1f-4273-ad90-21c08bf53c81 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Acquiring lock "5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.686628] env[63202]: DEBUG oslo_concurrency.lockutils [None req-50e06c6a-8e1f-4273-ad90-21c08bf53c81 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Lock "5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.686852] env[63202]: DEBUG oslo_concurrency.lockutils [None req-50e06c6a-8e1f-4273-ad90-21c08bf53c81 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Acquiring lock "5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.687165] env[63202]: DEBUG oslo_concurrency.lockutils [None req-50e06c6a-8e1f-4273-ad90-21c08bf53c81 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Lock "5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.687269] env[63202]: DEBUG oslo_concurrency.lockutils [None req-50e06c6a-8e1f-4273-ad90-21c08bf53c81 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Lock "5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.689662] env[63202]: INFO nova.compute.manager [None req-50e06c6a-8e1f-4273-ad90-21c08bf53c81 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Terminating instance [ 891.692145] env[63202]: DEBUG nova.compute.manager [None req-50e06c6a-8e1f-4273-ad90-21c08bf53c81 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 891.692541] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-50e06c6a-8e1f-4273-ad90-21c08bf53c81 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 891.693545] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6a88763-0261-4393-a23a-dc1ea0c55812 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.711839] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-50e06c6a-8e1f-4273-ad90-21c08bf53c81 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 891.714817] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7083a3bd-c9d8-4855-8419-10d222057861 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.716659] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4113c981-6ac7-436d-9548-9cdb3596054d tempest-ServersTestManualDisk-1560068381 tempest-ServersTestManualDisk-1560068381-project-member] Lock "ac48e21a-72ec-4c9d-8262-5d6cb36ec531" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 145.717s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.725039] env[63202]: DEBUG oslo_vmware.api [None req-50e06c6a-8e1f-4273-ad90-21c08bf53c81 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Waiting for the task: (returnval){ [ 891.725039] env[63202]: value = "task-1385420" [ 891.725039] env[63202]: _type = "Task" [ 891.725039] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.735222] env[63202]: DEBUG oslo_vmware.api [None req-50e06c6a-8e1f-4273-ad90-21c08bf53c81 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Task: {'id': task-1385420, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.755427] env[63202]: DEBUG nova.compute.manager [None req-1f54c5aa-21c2-4af5-9f98-bdb4568e182d tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] [instance: a800fa5d-46a1-4a96-8ea2-26e34b869aa6] Found 0 images (rotation: 2) {{(pid=63202) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 891.797017] env[63202]: DEBUG nova.compute.manager [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 891.820998] env[63202]: DEBUG oslo_vmware.api [None req-0915f74e-cfc6-45f1-9f44-9af3c69affef tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Task: {'id': task-1385418, 'name': PowerOffVM_Task, 'duration_secs': 0.1472} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.821313] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-0915f74e-cfc6-45f1-9f44-9af3c69affef tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] [instance: a800fa5d-46a1-4a96-8ea2-26e34b869aa6] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 891.821532] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-0915f74e-cfc6-45f1-9f44-9af3c69affef tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] [instance: a800fa5d-46a1-4a96-8ea2-26e34b869aa6] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 891.821754] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d990b0dc-dc85-41a1-9458-f63a1029a27a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.828902] env[63202]: DEBUG oslo_concurrency.lockutils [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.912s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.830018] env[63202]: ERROR nova.compute.manager [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 55d081d1-5a3b-4a11-9c16-b73b57c34ea3, please check neutron logs for more information. [ 891.830018] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] Traceback (most recent call last): [ 891.830018] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 891.830018] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] self.driver.spawn(context, instance, image_meta, [ 891.830018] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 891.830018] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 891.830018] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 891.830018] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] vm_ref = self.build_virtual_machine(instance, [ 891.830018] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 891.830018] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] vif_infos = vmwarevif.get_vif_info(self._session, [ 891.830018] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 891.830435] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] for vif in network_info: [ 891.830435] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 891.830435] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] return self._sync_wrapper(fn, *args, **kwargs) [ 891.830435] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 891.830435] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] self.wait() [ 891.830435] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 891.830435] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] self[:] = self._gt.wait() [ 891.830435] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 891.830435] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] return self._exit_event.wait() [ 891.830435] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 891.830435] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] result = hub.switch() [ 891.830435] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 891.830435] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] return self.greenlet.switch() [ 891.830789] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 891.830789] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] result = function(*args, **kwargs) [ 891.830789] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 891.830789] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] return func(*args, **kwargs) [ 891.830789] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 891.830789] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] raise e [ 891.830789] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 891.830789] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] nwinfo = self.network_api.allocate_for_instance( [ 891.830789] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 891.830789] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] created_port_ids = self._update_ports_for_instance( [ 891.830789] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 891.830789] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] with excutils.save_and_reraise_exception(): [ 891.830789] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 891.831181] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] self.force_reraise() [ 891.831181] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 891.831181] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] raise self.value [ 891.831181] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 891.831181] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] updated_port = self._update_port( [ 891.831181] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 891.831181] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] _ensure_no_port_binding_failure(port) [ 891.831181] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 891.831181] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] raise exception.PortBindingFailed(port_id=port['id']) [ 891.831181] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] nova.exception.PortBindingFailed: Binding failed for port 55d081d1-5a3b-4a11-9c16-b73b57c34ea3, please check neutron logs for more information. [ 891.831181] env[63202]: ERROR nova.compute.manager [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] [ 891.831544] env[63202]: DEBUG nova.compute.utils [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] Binding failed for port 55d081d1-5a3b-4a11-9c16-b73b57c34ea3, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 891.832629] env[63202]: DEBUG oslo_concurrency.lockutils [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.805s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.836960] env[63202]: DEBUG nova.compute.manager [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] Build of instance 5d9f1803-e3f8-42a5-b6b7-2642463a079e was re-scheduled: Binding failed for port 55d081d1-5a3b-4a11-9c16-b73b57c34ea3, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 891.838884] env[63202]: DEBUG nova.compute.manager [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 891.838884] env[63202]: DEBUG oslo_concurrency.lockutils [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Acquiring lock "refresh_cache-5d9f1803-e3f8-42a5-b6b7-2642463a079e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.838884] env[63202]: DEBUG oslo_concurrency.lockutils [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Acquired lock "refresh_cache-5d9f1803-e3f8-42a5-b6b7-2642463a079e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.838884] env[63202]: DEBUG nova.network.neutron [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 891.859078] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-0915f74e-cfc6-45f1-9f44-9af3c69affef tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] [instance: a800fa5d-46a1-4a96-8ea2-26e34b869aa6] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 891.859078] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-0915f74e-cfc6-45f1-9f44-9af3c69affef tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] [instance: a800fa5d-46a1-4a96-8ea2-26e34b869aa6] Deleting contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 891.859078] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-0915f74e-cfc6-45f1-9f44-9af3c69affef tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Deleting the datastore file [datastore1] a800fa5d-46a1-4a96-8ea2-26e34b869aa6 {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 891.859736] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-536ac88d-3299-483e-8866-adfea1d4d7fd {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.865735] env[63202]: DEBUG oslo_vmware.api [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385419, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.871243] env[63202]: DEBUG oslo_vmware.api [None req-0915f74e-cfc6-45f1-9f44-9af3c69affef tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Waiting for the task: (returnval){ [ 891.871243] env[63202]: value = "task-1385422" [ 891.871243] env[63202]: _type = "Task" [ 891.871243] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.881541] env[63202]: DEBUG oslo_vmware.api [None req-0915f74e-cfc6-45f1-9f44-9af3c69affef tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Task: {'id': task-1385422, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.013712] env[63202]: DEBUG oslo_concurrency.lockutils [req-fdd4414f-e7f0-482f-bc71-2aa538d7db0a req-b6a213a3-6de2-4c1d-8deb-08889df576ae service nova] Releasing lock "refresh_cache-2243bb69-0dc5-49cd-b94e-73e703cbadc0" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.013989] env[63202]: DEBUG nova.compute.manager [req-fdd4414f-e7f0-482f-bc71-2aa538d7db0a req-b6a213a3-6de2-4c1d-8deb-08889df576ae service nova] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Received event network-vif-deleted-f09247c0-c4a4-4583-95b6-54de5ec7b011 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 892.072388] env[63202]: INFO nova.compute.manager [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] [instance: 69c66a35-3789-4c2f-b7cf-5f4136d53b94] Took 1.04 seconds to deallocate network for instance. [ 892.074843] env[63202]: DEBUG oslo_vmware.api [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]527bd600-7ee3-2316-b548-caacf5423d07, 'name': SearchDatastore_Task, 'duration_secs': 0.039335} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.077622] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4bf98408-53cd-4471-86b7-638ea3edc71b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.083165] env[63202]: DEBUG oslo_vmware.api [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Waiting for the task: (returnval){ [ 892.083165] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52780f39-3259-dd77-bf3d-b48fe59e3882" [ 892.083165] env[63202]: _type = "Task" [ 892.083165] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.090734] env[63202]: DEBUG oslo_vmware.api [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52780f39-3259-dd77-bf3d-b48fe59e3882, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.219783] env[63202]: DEBUG nova.compute.manager [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 892.238426] env[63202]: DEBUG oslo_vmware.api [None req-50e06c6a-8e1f-4273-ad90-21c08bf53c81 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Task: {'id': task-1385420, 'name': PowerOffVM_Task, 'duration_secs': 0.243988} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.238796] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-50e06c6a-8e1f-4273-ad90-21c08bf53c81 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 892.239042] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-50e06c6a-8e1f-4273-ad90-21c08bf53c81 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 892.240038] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c2599ee5-cbe4-4e21-b2b4-43d0aad00250 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.319965] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-50e06c6a-8e1f-4273-ad90-21c08bf53c81 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 892.320234] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-50e06c6a-8e1f-4273-ad90-21c08bf53c81 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Deleting contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 892.320420] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-50e06c6a-8e1f-4273-ad90-21c08bf53c81 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Deleting the datastore file [datastore1] 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 892.320739] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-deea6d43-5409-4dd4-a3f7-f2ce3c441030 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.326537] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.332572] env[63202]: DEBUG oslo_vmware.api [None req-50e06c6a-8e1f-4273-ad90-21c08bf53c81 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Waiting for the task: (returnval){ [ 892.332572] env[63202]: value = "task-1385424" [ 892.332572] env[63202]: _type = "Task" [ 892.332572] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.351740] env[63202]: DEBUG oslo_vmware.api [None req-50e06c6a-8e1f-4273-ad90-21c08bf53c81 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Task: {'id': task-1385424, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.362188] env[63202]: DEBUG oslo_vmware.api [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385419, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.5289} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.364782] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore2] 50854a4e-8e4e-431b-8220-01bf8906b1c0/50854a4e-8e4e-431b-8220-01bf8906b1c0.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 892.365015] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 892.366291] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-255f1c77-a0ce-42c4-ad5e-0db9bcb13c9b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.382022] env[63202]: DEBUG oslo_vmware.api [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 892.382022] env[63202]: value = "task-1385425" [ 892.382022] env[63202]: _type = "Task" [ 892.382022] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.384143] env[63202]: DEBUG nova.network.neutron [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 892.389164] env[63202]: DEBUG oslo_vmware.api [None req-0915f74e-cfc6-45f1-9f44-9af3c69affef tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Task: {'id': task-1385422, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.397811] env[63202]: DEBUG oslo_vmware.api [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385425, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.502661] env[63202]: DEBUG nova.network.neutron [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.595314] env[63202]: DEBUG oslo_vmware.api [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52780f39-3259-dd77-bf3d-b48fe59e3882, 'name': SearchDatastore_Task, 'duration_secs': 0.038082} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.595558] env[63202]: DEBUG oslo_concurrency.lockutils [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.595810] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] 2243bb69-0dc5-49cd-b94e-73e703cbadc0/2243bb69-0dc5-49cd-b94e-73e703cbadc0.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 892.596076] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-20a12440-cfa8-4d96-970c-0e54fb3101cf {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.603799] env[63202]: DEBUG oslo_vmware.api [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Waiting for the task: (returnval){ [ 892.603799] env[63202]: value = "task-1385426" [ 892.603799] env[63202]: _type = "Task" [ 892.603799] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.611910] env[63202]: DEBUG oslo_vmware.api [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Task: {'id': task-1385426, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.636062] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d201eb65-327b-4439-9c02-dfd8c69624f7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.644310] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec6a0fdc-d5dd-425b-b7f8-a85df592a668 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.676432] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d433801-6b3c-4448-a8d7-9e232317e9cf {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.684883] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db3c703e-716f-47c2-9b89-a545de76b89b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.700389] env[63202]: DEBUG nova.compute.provider_tree [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 892.745650] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.842439] env[63202]: DEBUG oslo_vmware.api [None req-50e06c6a-8e1f-4273-ad90-21c08bf53c81 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Task: {'id': task-1385424, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.882165] env[63202]: DEBUG oslo_vmware.api [None req-0915f74e-cfc6-45f1-9f44-9af3c69affef tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Task: {'id': task-1385422, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.618586} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.885171] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-0915f74e-cfc6-45f1-9f44-9af3c69affef tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 892.885383] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-0915f74e-cfc6-45f1-9f44-9af3c69affef tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] [instance: a800fa5d-46a1-4a96-8ea2-26e34b869aa6] Deleted contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 892.885530] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-0915f74e-cfc6-45f1-9f44-9af3c69affef tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] [instance: a800fa5d-46a1-4a96-8ea2-26e34b869aa6] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 892.885694] env[63202]: INFO nova.compute.manager [None req-0915f74e-cfc6-45f1-9f44-9af3c69affef tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] [instance: a800fa5d-46a1-4a96-8ea2-26e34b869aa6] Took 1.60 seconds to destroy the instance on the hypervisor. [ 892.885919] env[63202]: DEBUG oslo.service.loopingcall [None req-0915f74e-cfc6-45f1-9f44-9af3c69affef tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 892.886151] env[63202]: DEBUG nova.compute.manager [-] [instance: a800fa5d-46a1-4a96-8ea2-26e34b869aa6] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 892.886240] env[63202]: DEBUG nova.network.neutron [-] [instance: a800fa5d-46a1-4a96-8ea2-26e34b869aa6] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 892.892996] env[63202]: DEBUG oslo_vmware.api [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385425, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082867} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.893258] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 892.894020] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57a78333-6658-4e86-94ad-a294c7651363 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.916452] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Reconfiguring VM instance instance-00000042 to attach disk [datastore2] 50854a4e-8e4e-431b-8220-01bf8906b1c0/50854a4e-8e4e-431b-8220-01bf8906b1c0.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 892.917351] env[63202]: DEBUG nova.network.neutron [-] [instance: a800fa5d-46a1-4a96-8ea2-26e34b869aa6] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 892.918536] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f144c01f-ac3d-4604-b1ad-e721c9a82c50 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.933521] env[63202]: DEBUG nova.network.neutron [-] [instance: a800fa5d-46a1-4a96-8ea2-26e34b869aa6] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.941589] env[63202]: DEBUG oslo_vmware.api [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 892.941589] env[63202]: value = "task-1385427" [ 892.941589] env[63202]: _type = "Task" [ 892.941589] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.951540] env[63202]: DEBUG oslo_vmware.api [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385427, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.008052] env[63202]: DEBUG oslo_concurrency.lockutils [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Releasing lock "refresh_cache-5d9f1803-e3f8-42a5-b6b7-2642463a079e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.008325] env[63202]: DEBUG nova.compute.manager [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 893.008548] env[63202]: DEBUG nova.compute.manager [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 893.008716] env[63202]: DEBUG nova.network.neutron [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 893.038388] env[63202]: DEBUG nova.network.neutron [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 893.104951] env[63202]: INFO nova.scheduler.client.report [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] Deleted allocations for instance 69c66a35-3789-4c2f-b7cf-5f4136d53b94 [ 893.121247] env[63202]: DEBUG oslo_vmware.api [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Task: {'id': task-1385426, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.203957] env[63202]: DEBUG nova.scheduler.client.report [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 893.343374] env[63202]: DEBUG oslo_vmware.api [None req-50e06c6a-8e1f-4273-ad90-21c08bf53c81 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Task: {'id': task-1385424, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.437598] env[63202]: INFO nova.compute.manager [-] [instance: a800fa5d-46a1-4a96-8ea2-26e34b869aa6] Took 0.55 seconds to deallocate network for instance. [ 893.459720] env[63202]: DEBUG oslo_vmware.api [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385427, 'name': ReconfigVM_Task, 'duration_secs': 0.340192} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.460165] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Reconfigured VM instance instance-00000042 to attach disk [datastore2] 50854a4e-8e4e-431b-8220-01bf8906b1c0/50854a4e-8e4e-431b-8220-01bf8906b1c0.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 893.461029] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-613d4c53-f17f-4f7e-96ba-09eae6370f40 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.475939] env[63202]: DEBUG oslo_vmware.api [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 893.475939] env[63202]: value = "task-1385428" [ 893.475939] env[63202]: _type = "Task" [ 893.475939] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.490300] env[63202]: DEBUG oslo_vmware.api [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385428, 'name': Rename_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.541984] env[63202]: DEBUG nova.network.neutron [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 893.616207] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1537a42a-eb97-42f3-8eba-21728d271e76 tempest-ServerMetadataNegativeTestJSON-87994245 tempest-ServerMetadataNegativeTestJSON-87994245-project-member] Lock "69c66a35-3789-4c2f-b7cf-5f4136d53b94" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 144.190s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.639778] env[63202]: DEBUG oslo_vmware.api [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Task: {'id': task-1385426, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.712955] env[63202]: DEBUG oslo_concurrency.lockutils [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.880s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.713653] env[63202]: ERROR nova.compute.manager [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: d674db7a-8cf0-46f8-b676-fced0767c105] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8852a0b8-5683-4012-ae86-2dde3afc9dbf, please check neutron logs for more information. [ 893.713653] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] Traceback (most recent call last): [ 893.713653] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 893.713653] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] self.driver.spawn(context, instance, image_meta, [ 893.713653] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 893.713653] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] self._vmops.spawn(context, instance, image_meta, injected_files, [ 893.713653] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 893.713653] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] vm_ref = self.build_virtual_machine(instance, [ 893.713653] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 893.713653] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] vif_infos = vmwarevif.get_vif_info(self._session, [ 893.713653] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 893.714112] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] for vif in network_info: [ 893.714112] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 893.714112] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] return self._sync_wrapper(fn, *args, **kwargs) [ 893.714112] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 893.714112] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] self.wait() [ 893.714112] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 893.714112] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] self[:] = self._gt.wait() [ 893.714112] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 893.714112] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] return self._exit_event.wait() [ 893.714112] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 893.714112] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] current.throw(*self._exc) [ 893.714112] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 893.714112] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] result = function(*args, **kwargs) [ 893.714496] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 893.714496] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] return func(*args, **kwargs) [ 893.714496] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 893.714496] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] raise e [ 893.714496] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 893.714496] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] nwinfo = self.network_api.allocate_for_instance( [ 893.714496] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 893.714496] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] created_port_ids = self._update_ports_for_instance( [ 893.714496] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 893.714496] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] with excutils.save_and_reraise_exception(): [ 893.714496] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 893.714496] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] self.force_reraise() [ 893.714496] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 893.714830] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] raise self.value [ 893.714830] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 893.714830] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] updated_port = self._update_port( [ 893.714830] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 893.714830] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] _ensure_no_port_binding_failure(port) [ 893.714830] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 893.714830] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] raise exception.PortBindingFailed(port_id=port['id']) [ 893.714830] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] nova.exception.PortBindingFailed: Binding failed for port 8852a0b8-5683-4012-ae86-2dde3afc9dbf, please check neutron logs for more information. [ 893.714830] env[63202]: ERROR nova.compute.manager [instance: d674db7a-8cf0-46f8-b676-fced0767c105] [ 893.714830] env[63202]: DEBUG nova.compute.utils [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: d674db7a-8cf0-46f8-b676-fced0767c105] Binding failed for port 8852a0b8-5683-4012-ae86-2dde3afc9dbf, please check neutron logs for more information. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 893.716556] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 9.959s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.716556] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.716556] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63202) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 893.716556] env[63202]: DEBUG oslo_concurrency.lockutils [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.260s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.718386] env[63202]: INFO nova.compute.claims [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 893.721901] env[63202]: DEBUG nova.compute.manager [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: d674db7a-8cf0-46f8-b676-fced0767c105] Build of instance d674db7a-8cf0-46f8-b676-fced0767c105 was re-scheduled: Binding failed for port 8852a0b8-5683-4012-ae86-2dde3afc9dbf, please check neutron logs for more information. {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 893.722430] env[63202]: DEBUG nova.compute.manager [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: d674db7a-8cf0-46f8-b676-fced0767c105] Unplugging VIFs for instance {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 893.722660] env[63202]: DEBUG oslo_concurrency.lockutils [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "refresh_cache-d674db7a-8cf0-46f8-b676-fced0767c105" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.722804] env[63202]: DEBUG oslo_concurrency.lockutils [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquired lock "refresh_cache-d674db7a-8cf0-46f8-b676-fced0767c105" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.722960] env[63202]: DEBUG nova.network.neutron [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: d674db7a-8cf0-46f8-b676-fced0767c105] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 893.724558] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3cd0835-9e5e-4a88-ada9-6137e2636c5d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.735302] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-647229f7-793b-4203-ba3b-98bda34ab34d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.752411] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c22c8458-08d3-4dc4-8e2e-d9870825e6ac {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.759932] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8345498-e2a4-4038-93dc-b29544f62883 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.791795] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181324MB free_disk=176GB free_vcpus=48 pci_devices=None {{(pid=63202) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 893.791952] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.842665] env[63202]: DEBUG oslo_vmware.api [None req-50e06c6a-8e1f-4273-ad90-21c08bf53c81 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Task: {'id': task-1385424, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.049791} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.842906] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-50e06c6a-8e1f-4273-ad90-21c08bf53c81 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 893.843103] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-50e06c6a-8e1f-4273-ad90-21c08bf53c81 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Deleted contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 893.843273] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-50e06c6a-8e1f-4273-ad90-21c08bf53c81 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 893.843437] env[63202]: INFO nova.compute.manager [None req-50e06c6a-8e1f-4273-ad90-21c08bf53c81 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Took 2.15 seconds to destroy the instance on the hypervisor. [ 893.843667] env[63202]: DEBUG oslo.service.loopingcall [None req-50e06c6a-8e1f-4273-ad90-21c08bf53c81 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 893.843849] env[63202]: DEBUG nova.compute.manager [-] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 893.843939] env[63202]: DEBUG nova.network.neutron [-] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 893.946917] env[63202]: DEBUG oslo_concurrency.lockutils [None req-0915f74e-cfc6-45f1-9f44-9af3c69affef tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.985184] env[63202]: DEBUG oslo_vmware.api [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385428, 'name': Rename_Task, 'duration_secs': 0.231173} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.985495] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 893.985741] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c0ba3e53-5682-4716-a2c0-f81020303bfd {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.992866] env[63202]: DEBUG oslo_vmware.api [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 893.992866] env[63202]: value = "task-1385429" [ 893.992866] env[63202]: _type = "Task" [ 893.992866] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.003471] env[63202]: DEBUG oslo_vmware.api [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385429, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.046686] env[63202]: INFO nova.compute.manager [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] [instance: 5d9f1803-e3f8-42a5-b6b7-2642463a079e] Took 1.04 seconds to deallocate network for instance. [ 894.107116] env[63202]: DEBUG nova.compute.manager [req-3de89ddf-5245-496e-9cf2-3035f8467098 req-51e72821-72c8-46ee-b70e-c0608fa557c2 service nova] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Received event network-vif-deleted-d12e501a-1681-4490-84f1-f48efcf89671 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 894.107480] env[63202]: INFO nova.compute.manager [req-3de89ddf-5245-496e-9cf2-3035f8467098 req-51e72821-72c8-46ee-b70e-c0608fa557c2 service nova] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Neutron deleted interface d12e501a-1681-4490-84f1-f48efcf89671; detaching it from the instance and deleting it from the info cache [ 894.107662] env[63202]: DEBUG nova.network.neutron [req-3de89ddf-5245-496e-9cf2-3035f8467098 req-51e72821-72c8-46ee-b70e-c0608fa557c2 service nova] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.119734] env[63202]: DEBUG nova.compute.manager [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 894.126208] env[63202]: DEBUG oslo_vmware.api [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Task: {'id': task-1385426, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.154937} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.126528] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] 2243bb69-0dc5-49cd-b94e-73e703cbadc0/2243bb69-0dc5-49cd-b94e-73e703cbadc0.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 894.126741] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 894.127237] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-90246b12-a6b3-4ebe-b82f-68df45cdb366 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.135917] env[63202]: DEBUG oslo_vmware.api [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Waiting for the task: (returnval){ [ 894.135917] env[63202]: value = "task-1385430" [ 894.135917] env[63202]: _type = "Task" [ 894.135917] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.145504] env[63202]: DEBUG oslo_vmware.api [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Task: {'id': task-1385430, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.248227] env[63202]: DEBUG nova.network.neutron [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: d674db7a-8cf0-46f8-b676-fced0767c105] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 894.355140] env[63202]: DEBUG nova.network.neutron [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: d674db7a-8cf0-46f8-b676-fced0767c105] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.504507] env[63202]: DEBUG oslo_vmware.api [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385429, 'name': PowerOnVM_Task, 'duration_secs': 0.480617} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.504701] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 894.504891] env[63202]: INFO nova.compute.manager [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Took 8.90 seconds to spawn the instance on the hypervisor. [ 894.505080] env[63202]: DEBUG nova.compute.manager [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 894.505861] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13007870-1609-4160-9bc6-6b71c72b84f1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.587037] env[63202]: DEBUG nova.network.neutron [-] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.611979] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6ae31a08-96ce-4c85-a7df-d8eac6051f56 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.623574] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a8dcfe9-a4c9-426e-98a6-76b470388b59 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.647327] env[63202]: DEBUG oslo_vmware.api [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Task: {'id': task-1385430, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064704} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.648054] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 894.648401] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-138fe08f-07d4-4e08-8fda-e5d04e9d2621 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.658262] env[63202]: DEBUG nova.compute.manager [req-3de89ddf-5245-496e-9cf2-3035f8467098 req-51e72821-72c8-46ee-b70e-c0608fa557c2 service nova] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Detach interface failed, port_id=d12e501a-1681-4490-84f1-f48efcf89671, reason: Instance 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a could not be found. {{(pid=63202) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 894.659440] env[63202]: DEBUG oslo_concurrency.lockutils [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.680632] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] 2243bb69-0dc5-49cd-b94e-73e703cbadc0/2243bb69-0dc5-49cd-b94e-73e703cbadc0.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 894.680965] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-94488363-92bc-4ac8-9560-74f0acb4de27 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.704090] env[63202]: DEBUG oslo_vmware.api [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Waiting for the task: (returnval){ [ 894.704090] env[63202]: value = "task-1385431" [ 894.704090] env[63202]: _type = "Task" [ 894.704090] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.714303] env[63202]: DEBUG oslo_vmware.api [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Task: {'id': task-1385431, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.858154] env[63202]: DEBUG oslo_concurrency.lockutils [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Releasing lock "refresh_cache-d674db7a-8cf0-46f8-b676-fced0767c105" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.858470] env[63202]: DEBUG nova.compute.manager [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63202) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 894.858657] env[63202]: DEBUG nova.compute.manager [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: d674db7a-8cf0-46f8-b676-fced0767c105] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 894.858841] env[63202]: DEBUG nova.network.neutron [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: d674db7a-8cf0-46f8-b676-fced0767c105] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 894.887778] env[63202]: DEBUG nova.network.neutron [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: d674db7a-8cf0-46f8-b676-fced0767c105] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 895.026945] env[63202]: INFO nova.compute.manager [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Took 41.81 seconds to build instance. [ 895.055544] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb16f0ae-aede-47aa-8440-cb1b4a9f700e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.074253] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a1f839b-3f69-4bae-8088-b199329feebb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.083041] env[63202]: INFO nova.scheduler.client.report [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Deleted allocations for instance 5d9f1803-e3f8-42a5-b6b7-2642463a079e [ 895.116232] env[63202]: INFO nova.compute.manager [-] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Took 1.27 seconds to deallocate network for instance. [ 895.121584] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c32c95a8-0235-44b2-87ab-00568806fcc2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.133930] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b12d0e02-be58-4006-b551-f19255179cd9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.154813] env[63202]: DEBUG nova.compute.provider_tree [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 895.215380] env[63202]: DEBUG oslo_vmware.api [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Task: {'id': task-1385431, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.393493] env[63202]: DEBUG nova.network.neutron [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: d674db7a-8cf0-46f8-b676-fced0767c105] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.531731] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ba2328f9-5a46-40ae-8ad0-727f8d7b5494 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "50854a4e-8e4e-431b-8220-01bf8906b1c0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 98.185s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.619897] env[63202]: DEBUG oslo_concurrency.lockutils [None req-49fce254-1c62-4e5a-b525-93689db334b2 tempest-VolumesAdminNegativeTest-1985833072 tempest-VolumesAdminNegativeTest-1985833072-project-member] Lock "5d9f1803-e3f8-42a5-b6b7-2642463a079e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 144.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.629701] env[63202]: DEBUG oslo_concurrency.lockutils [None req-50e06c6a-8e1f-4273-ad90-21c08bf53c81 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.659150] env[63202]: DEBUG nova.scheduler.client.report [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 895.718301] env[63202]: DEBUG oslo_vmware.api [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Task: {'id': task-1385431, 'name': ReconfigVM_Task, 'duration_secs': 0.857692} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.718919] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Reconfigured VM instance instance-00000043 to attach disk [datastore1] 2243bb69-0dc5-49cd-b94e-73e703cbadc0/2243bb69-0dc5-49cd-b94e-73e703cbadc0.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 895.719379] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ab0b4ca0-28af-4247-8d0a-71212f810dc1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.728306] env[63202]: DEBUG oslo_vmware.api [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Waiting for the task: (returnval){ [ 895.728306] env[63202]: value = "task-1385432" [ 895.728306] env[63202]: _type = "Task" [ 895.728306] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.739668] env[63202]: DEBUG oslo_vmware.api [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Task: {'id': task-1385432, 'name': Rename_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.897343] env[63202]: INFO nova.compute.manager [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: d674db7a-8cf0-46f8-b676-fced0767c105] Took 1.04 seconds to deallocate network for instance. [ 896.036464] env[63202]: DEBUG nova.compute.manager [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 896.123595] env[63202]: DEBUG nova.compute.manager [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 896.166038] env[63202]: DEBUG oslo_concurrency.lockutils [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.449s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.166589] env[63202]: DEBUG nova.compute.manager [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 896.169656] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.555s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.171185] env[63202]: INFO nova.compute.claims [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 896.241288] env[63202]: DEBUG oslo_vmware.api [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Task: {'id': task-1385432, 'name': Rename_Task, 'duration_secs': 0.159057} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.241288] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 896.241288] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b888a8ec-a54c-4f9e-ae3e-f2a47b117a5a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.248317] env[63202]: DEBUG oslo_vmware.api [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Waiting for the task: (returnval){ [ 896.248317] env[63202]: value = "task-1385433" [ 896.248317] env[63202]: _type = "Task" [ 896.248317] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.259721] env[63202]: DEBUG oslo_vmware.api [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Task: {'id': task-1385433, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.561769] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.646100] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.647600] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f347bd6c-7136-4427-b9a2-12f6c6163431 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "50854a4e-8e4e-431b-8220-01bf8906b1c0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.647841] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f347bd6c-7136-4427-b9a2-12f6c6163431 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "50854a4e-8e4e-431b-8220-01bf8906b1c0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.648051] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f347bd6c-7136-4427-b9a2-12f6c6163431 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "50854a4e-8e4e-431b-8220-01bf8906b1c0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.648236] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f347bd6c-7136-4427-b9a2-12f6c6163431 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "50854a4e-8e4e-431b-8220-01bf8906b1c0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.649024] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f347bd6c-7136-4427-b9a2-12f6c6163431 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "50854a4e-8e4e-431b-8220-01bf8906b1c0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.651905] env[63202]: INFO nova.compute.manager [None req-f347bd6c-7136-4427-b9a2-12f6c6163431 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Terminating instance [ 896.654275] env[63202]: DEBUG nova.compute.manager [None req-f347bd6c-7136-4427-b9a2-12f6c6163431 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 896.656415] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-f347bd6c-7136-4427-b9a2-12f6c6163431 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 896.656415] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f90999d7-50f5-4837-ae6e-83c91de30bf0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.665090] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-f347bd6c-7136-4427-b9a2-12f6c6163431 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 896.665363] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-614b6ee0-12c4-42f2-9603-57816e2c0f13 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.676112] env[63202]: DEBUG nova.compute.utils [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 896.681879] env[63202]: DEBUG nova.compute.manager [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 896.681879] env[63202]: DEBUG nova.network.neutron [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 896.682318] env[63202]: DEBUG oslo_vmware.api [None req-f347bd6c-7136-4427-b9a2-12f6c6163431 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 896.682318] env[63202]: value = "task-1385434" [ 896.682318] env[63202]: _type = "Task" [ 896.682318] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.692305] env[63202]: DEBUG oslo_vmware.api [None req-f347bd6c-7136-4427-b9a2-12f6c6163431 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385434, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.722431] env[63202]: DEBUG nova.policy [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'faabe4f6570d41598fd1f2b2f0d6df11', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '98fbc2f9559148b8b0e2a2136152f13c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 896.763148] env[63202]: DEBUG oslo_vmware.api [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Task: {'id': task-1385433, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.936022] env[63202]: INFO nova.scheduler.client.report [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Deleted allocations for instance d674db7a-8cf0-46f8-b676-fced0767c105 [ 897.048784] env[63202]: DEBUG nova.network.neutron [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] Successfully created port: 355608b4-3a2c-4f49-8b9a-6b379a97934d {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 897.184124] env[63202]: DEBUG nova.compute.manager [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 897.206742] env[63202]: DEBUG oslo_vmware.api [None req-f347bd6c-7136-4427-b9a2-12f6c6163431 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385434, 'name': PowerOffVM_Task, 'duration_secs': 0.191719} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.207686] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-f347bd6c-7136-4427-b9a2-12f6c6163431 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 897.207915] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-f347bd6c-7136-4427-b9a2-12f6c6163431 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 897.208233] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-61339964-66bc-4b10-886c-549a71f62c57 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.265514] env[63202]: DEBUG oslo_vmware.api [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Task: {'id': task-1385433, 'name': PowerOnVM_Task, 'duration_secs': 0.817562} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.266127] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 897.266377] env[63202]: INFO nova.compute.manager [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Took 9.19 seconds to spawn the instance on the hypervisor. [ 897.266585] env[63202]: DEBUG nova.compute.manager [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 897.267450] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dd99c8f-4e48-4b25-9eb0-84d9920a79b0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.276673] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-f347bd6c-7136-4427-b9a2-12f6c6163431 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 897.276891] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-f347bd6c-7136-4427-b9a2-12f6c6163431 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Deleting contents of the VM from datastore datastore2 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 897.277084] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-f347bd6c-7136-4427-b9a2-12f6c6163431 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Deleting the datastore file [datastore2] 50854a4e-8e4e-431b-8220-01bf8906b1c0 {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 897.277348] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-71498040-7355-461b-817d-25345b944d60 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.288705] env[63202]: DEBUG oslo_vmware.api [None req-f347bd6c-7136-4427-b9a2-12f6c6163431 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 897.288705] env[63202]: value = "task-1385436" [ 897.288705] env[63202]: _type = "Task" [ 897.288705] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.302709] env[63202]: DEBUG oslo_vmware.api [None req-f347bd6c-7136-4427-b9a2-12f6c6163431 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385436, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.446858] env[63202]: DEBUG oslo_concurrency.lockutils [None req-32e8198d-f13d-499e-86a9-a0df59433cc0 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "d674db7a-8cf0-46f8-b676-fced0767c105" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 137.054s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.510528] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21718e60-1f22-4b70-abdc-cdcf262e2cf8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.520112] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be91619c-7691-4df6-aea1-17bf9c10166d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.557180] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba132baa-a489-4649-9296-793c404442ac {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.569596] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d234a3c2-6165-4517-a5d3-96db65577a24 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.585582] env[63202]: DEBUG nova.compute.provider_tree [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 897.797140] env[63202]: INFO nova.compute.manager [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Took 42.50 seconds to build instance. [ 897.808456] env[63202]: DEBUG oslo_vmware.api [None req-f347bd6c-7136-4427-b9a2-12f6c6163431 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385436, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.349398} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.808751] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-f347bd6c-7136-4427-b9a2-12f6c6163431 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 897.808844] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-f347bd6c-7136-4427-b9a2-12f6c6163431 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Deleted contents of the VM from datastore datastore2 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 897.809137] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-f347bd6c-7136-4427-b9a2-12f6c6163431 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 897.809319] env[63202]: INFO nova.compute.manager [None req-f347bd6c-7136-4427-b9a2-12f6c6163431 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Took 1.15 seconds to destroy the instance on the hypervisor. [ 897.810046] env[63202]: DEBUG oslo.service.loopingcall [None req-f347bd6c-7136-4427-b9a2-12f6c6163431 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 897.810046] env[63202]: DEBUG nova.compute.manager [-] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 897.810046] env[63202]: DEBUG nova.network.neutron [-] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 897.955017] env[63202]: DEBUG nova.compute.manager [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 898.090195] env[63202]: DEBUG nova.scheduler.client.report [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 898.207237] env[63202]: DEBUG nova.compute.manager [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 898.239957] env[63202]: DEBUG nova.virt.hardware [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 898.240883] env[63202]: DEBUG nova.virt.hardware [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 898.241117] env[63202]: DEBUG nova.virt.hardware [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 898.241324] env[63202]: DEBUG nova.virt.hardware [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 898.241449] env[63202]: DEBUG nova.virt.hardware [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 898.241596] env[63202]: DEBUG nova.virt.hardware [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 898.241798] env[63202]: DEBUG nova.virt.hardware [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 898.242056] env[63202]: DEBUG nova.virt.hardware [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 898.242137] env[63202]: DEBUG nova.virt.hardware [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 898.242368] env[63202]: DEBUG nova.virt.hardware [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 898.242528] env[63202]: DEBUG nova.virt.hardware [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 898.243342] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0463334-250c-4391-8745-325eeab8722f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.247933] env[63202]: DEBUG nova.compute.manager [req-6b72a7cb-b148-4342-9ed2-86ff7117e1fc req-213a2fb4-5dd4-462f-8684-574bf3fd3ec1 service nova] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Received event network-vif-deleted-6061b268-0c4c-414c-a0a3-a60196071be2 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 898.247986] env[63202]: INFO nova.compute.manager [req-6b72a7cb-b148-4342-9ed2-86ff7117e1fc req-213a2fb4-5dd4-462f-8684-574bf3fd3ec1 service nova] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Neutron deleted interface 6061b268-0c4c-414c-a0a3-a60196071be2; detaching it from the instance and deleting it from the info cache [ 898.248134] env[63202]: DEBUG nova.network.neutron [req-6b72a7cb-b148-4342-9ed2-86ff7117e1fc req-213a2fb4-5dd4-462f-8684-574bf3fd3ec1 service nova] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 898.259117] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fd2ea61-8129-4cea-93f3-493886b6dc41 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.301910] env[63202]: DEBUG oslo_concurrency.lockutils [None req-61740d7d-40c9-4a37-a20b-afe2628c810b tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Lock "2243bb69-0dc5-49cd-b94e-73e703cbadc0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 95.251s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.489582] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.595890] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.426s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.596498] env[63202]: DEBUG nova.compute.manager [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 898.599127] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7a451f85-aa73-49f0-af5e-6452e182ee9b tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.094s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.599470] env[63202]: DEBUG nova.objects.instance [None req-7a451f85-aa73-49f0-af5e-6452e182ee9b tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Lazy-loading 'resources' on Instance uuid cedc3a06-2123-4c5e-a6c3-599a3efc3c65 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 898.645069] env[63202]: DEBUG nova.network.neutron [-] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 898.750790] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1ab04da3-8b78-4d2a-b51c-3e10c66975f8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.760886] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-703ce62e-3445-43e6-ad9a-4f4f6d295d84 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.791020] env[63202]: DEBUG nova.compute.manager [req-6b72a7cb-b148-4342-9ed2-86ff7117e1fc req-213a2fb4-5dd4-462f-8684-574bf3fd3ec1 service nova] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Detach interface failed, port_id=6061b268-0c4c-414c-a0a3-a60196071be2, reason: Instance 50854a4e-8e4e-431b-8220-01bf8906b1c0 could not be found. {{(pid=63202) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 898.805125] env[63202]: DEBUG nova.compute.manager [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 898.894023] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "8755bceb-d510-4429-bd98-d6a63faf739d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.894164] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "8755bceb-d510-4429-bd98-d6a63faf739d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.060862] env[63202]: DEBUG nova.compute.manager [req-4a515706-a96b-4b12-878e-98d98246a32d req-6604708b-de1c-43a9-9ab8-cd245dfb8c82 service nova] [instance: 36291628-af7d-43cf-8149-09f57df47890] Received event network-vif-plugged-355608b4-3a2c-4f49-8b9a-6b379a97934d {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 899.060993] env[63202]: DEBUG oslo_concurrency.lockutils [req-4a515706-a96b-4b12-878e-98d98246a32d req-6604708b-de1c-43a9-9ab8-cd245dfb8c82 service nova] Acquiring lock "36291628-af7d-43cf-8149-09f57df47890-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.061226] env[63202]: DEBUG oslo_concurrency.lockutils [req-4a515706-a96b-4b12-878e-98d98246a32d req-6604708b-de1c-43a9-9ab8-cd245dfb8c82 service nova] Lock "36291628-af7d-43cf-8149-09f57df47890-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.061398] env[63202]: DEBUG oslo_concurrency.lockutils [req-4a515706-a96b-4b12-878e-98d98246a32d req-6604708b-de1c-43a9-9ab8-cd245dfb8c82 service nova] Lock "36291628-af7d-43cf-8149-09f57df47890-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.061563] env[63202]: DEBUG nova.compute.manager [req-4a515706-a96b-4b12-878e-98d98246a32d req-6604708b-de1c-43a9-9ab8-cd245dfb8c82 service nova] [instance: 36291628-af7d-43cf-8149-09f57df47890] No waiting events found dispatching network-vif-plugged-355608b4-3a2c-4f49-8b9a-6b379a97934d {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 899.061753] env[63202]: WARNING nova.compute.manager [req-4a515706-a96b-4b12-878e-98d98246a32d req-6604708b-de1c-43a9-9ab8-cd245dfb8c82 service nova] [instance: 36291628-af7d-43cf-8149-09f57df47890] Received unexpected event network-vif-plugged-355608b4-3a2c-4f49-8b9a-6b379a97934d for instance with vm_state building and task_state spawning. [ 899.105903] env[63202]: DEBUG nova.compute.utils [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 899.107791] env[63202]: DEBUG nova.compute.manager [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 899.111141] env[63202]: DEBUG nova.network.neutron [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 899.147660] env[63202]: INFO nova.compute.manager [-] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Took 1.34 seconds to deallocate network for instance. [ 899.185947] env[63202]: DEBUG nova.policy [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'faabe4f6570d41598fd1f2b2f0d6df11', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '98fbc2f9559148b8b0e2a2136152f13c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 899.244858] env[63202]: DEBUG nova.network.neutron [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] Successfully updated port: 355608b4-3a2c-4f49-8b9a-6b379a97934d {{(pid=63202) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 899.337805] env[63202]: DEBUG oslo_concurrency.lockutils [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.444751] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ba6e826-2b77-4b51-8f5d-b8c165e97fc7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.455071] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75f69947-c151-454e-8bd9-ae5bd8476a12 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.494762] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-887171df-05bf-4694-9b75-469ffc7b7ace {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.505443] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0839f0d6-ae1a-4942-a004-24f23599a427 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.520127] env[63202]: DEBUG nova.compute.provider_tree [None req-7a451f85-aa73-49f0-af5e-6452e182ee9b tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 899.613690] env[63202]: DEBUG nova.compute.manager [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 899.651078] env[63202]: DEBUG nova.network.neutron [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Successfully created port: 06e55a71-f982-49f8-819f-3408e19025de {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 899.658488] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f347bd6c-7136-4427-b9a2-12f6c6163431 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.752833] env[63202]: DEBUG oslo_concurrency.lockutils [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Acquiring lock "refresh_cache-36291628-af7d-43cf-8149-09f57df47890" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 899.756176] env[63202]: DEBUG oslo_concurrency.lockutils [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Acquired lock "refresh_cache-36291628-af7d-43cf-8149-09f57df47890" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.756176] env[63202]: DEBUG nova.network.neutron [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 900.024930] env[63202]: DEBUG nova.scheduler.client.report [None req-7a451f85-aa73-49f0-af5e-6452e182ee9b tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 900.308845] env[63202]: DEBUG nova.network.neutron [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 900.488266] env[63202]: DEBUG nova.network.neutron [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] Updating instance_info_cache with network_info: [{"id": "355608b4-3a2c-4f49-8b9a-6b379a97934d", "address": "fa:16:3e:59:3b:cb", "network": {"id": "d83c3dc7-a0c9-4401-8b08-99d493f23a8b", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-482332878-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "98fbc2f9559148b8b0e2a2136152f13c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap355608b4-3a", "ovs_interfaceid": "355608b4-3a2c-4f49-8b9a-6b379a97934d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.531882] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7a451f85-aa73-49f0-af5e-6452e182ee9b tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.933s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.534183] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.803s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.535657] env[63202]: INFO nova.compute.claims [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 900.558521] env[63202]: INFO nova.scheduler.client.report [None req-7a451f85-aa73-49f0-af5e-6452e182ee9b tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Deleted allocations for instance cedc3a06-2123-4c5e-a6c3-599a3efc3c65 [ 900.628244] env[63202]: DEBUG nova.compute.manager [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 900.659021] env[63202]: DEBUG nova.virt.hardware [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 900.659369] env[63202]: DEBUG nova.virt.hardware [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 900.660371] env[63202]: DEBUG nova.virt.hardware [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 900.660371] env[63202]: DEBUG nova.virt.hardware [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 900.660371] env[63202]: DEBUG nova.virt.hardware [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 900.660371] env[63202]: DEBUG nova.virt.hardware [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 900.660371] env[63202]: DEBUG nova.virt.hardware [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 900.660567] env[63202]: DEBUG nova.virt.hardware [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 900.660743] env[63202]: DEBUG nova.virt.hardware [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 900.660912] env[63202]: DEBUG nova.virt.hardware [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 900.661096] env[63202]: DEBUG nova.virt.hardware [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 900.662039] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b0b9aa5-2805-4f3f-96b1-b944adc38a4b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.671899] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-847b2f62-5329-4cfa-84ee-5a08951ec340 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.993356] env[63202]: DEBUG oslo_concurrency.lockutils [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Releasing lock "refresh_cache-36291628-af7d-43cf-8149-09f57df47890" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.993356] env[63202]: DEBUG nova.compute.manager [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] Instance network_info: |[{"id": "355608b4-3a2c-4f49-8b9a-6b379a97934d", "address": "fa:16:3e:59:3b:cb", "network": {"id": "d83c3dc7-a0c9-4401-8b08-99d493f23a8b", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-482332878-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "98fbc2f9559148b8b0e2a2136152f13c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap355608b4-3a", "ovs_interfaceid": "355608b4-3a2c-4f49-8b9a-6b379a97934d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 900.993561] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:59:3b:cb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f3ccbdbb-8b49-4a26-913f-2a448b72280f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '355608b4-3a2c-4f49-8b9a-6b379a97934d', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 901.004877] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Creating folder: Project (98fbc2f9559148b8b0e2a2136152f13c). Parent ref: group-v294090. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 901.004877] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a35c387c-65c7-4c1c-921b-041800fad751 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.020551] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Created folder: Project (98fbc2f9559148b8b0e2a2136152f13c) in parent group-v294090. [ 901.020755] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Creating folder: Instances. Parent ref: group-v294130. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 901.021020] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-06c2fb66-1fe0-4685-ab1a-5b17838d7f70 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.035368] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Created folder: Instances in parent group-v294130. [ 901.035625] env[63202]: DEBUG oslo.service.loopingcall [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 901.035821] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 36291628-af7d-43cf-8149-09f57df47890] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 901.036035] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e42693fd-7ec2-42e9-bf1b-e6e7fc46c53a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.067224] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7a451f85-aa73-49f0-af5e-6452e182ee9b tempest-ServerMetadataTestJSON-985543558 tempest-ServerMetadataTestJSON-985543558-project-member] Lock "cedc3a06-2123-4c5e-a6c3-599a3efc3c65" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.514s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.068336] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 901.068336] env[63202]: value = "task-1385439" [ 901.068336] env[63202]: _type = "Task" [ 901.068336] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.082478] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385439, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.137032] env[63202]: DEBUG nova.compute.manager [req-c212a235-4621-4050-9c13-2533dd9298fd req-1c07f22b-c89e-40f1-aea7-f7e8f58dfe7b service nova] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Received event network-changed-a0175615-5853-4224-8504-2a296446e39b {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 901.137442] env[63202]: DEBUG nova.compute.manager [req-c212a235-4621-4050-9c13-2533dd9298fd req-1c07f22b-c89e-40f1-aea7-f7e8f58dfe7b service nova] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Refreshing instance network info cache due to event network-changed-a0175615-5853-4224-8504-2a296446e39b. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 901.137801] env[63202]: DEBUG oslo_concurrency.lockutils [req-c212a235-4621-4050-9c13-2533dd9298fd req-1c07f22b-c89e-40f1-aea7-f7e8f58dfe7b service nova] Acquiring lock "refresh_cache-2243bb69-0dc5-49cd-b94e-73e703cbadc0" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.137869] env[63202]: DEBUG oslo_concurrency.lockutils [req-c212a235-4621-4050-9c13-2533dd9298fd req-1c07f22b-c89e-40f1-aea7-f7e8f58dfe7b service nova] Acquired lock "refresh_cache-2243bb69-0dc5-49cd-b94e-73e703cbadc0" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.138479] env[63202]: DEBUG nova.network.neutron [req-c212a235-4621-4050-9c13-2533dd9298fd req-1c07f22b-c89e-40f1-aea7-f7e8f58dfe7b service nova] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Refreshing network info cache for port a0175615-5853-4224-8504-2a296446e39b {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 901.388024] env[63202]: DEBUG nova.compute.manager [req-7740c387-a001-4c3a-9efe-5ed830df2640 req-73e2c303-fcf0-40e6-b306-db1b35a1807d service nova] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Received event network-vif-plugged-06e55a71-f982-49f8-819f-3408e19025de {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 901.388249] env[63202]: DEBUG oslo_concurrency.lockutils [req-7740c387-a001-4c3a-9efe-5ed830df2640 req-73e2c303-fcf0-40e6-b306-db1b35a1807d service nova] Acquiring lock "3d1518d3-bed2-4b7b-af1d-d7da49e92874-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.388452] env[63202]: DEBUG oslo_concurrency.lockutils [req-7740c387-a001-4c3a-9efe-5ed830df2640 req-73e2c303-fcf0-40e6-b306-db1b35a1807d service nova] Lock "3d1518d3-bed2-4b7b-af1d-d7da49e92874-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.388615] env[63202]: DEBUG oslo_concurrency.lockutils [req-7740c387-a001-4c3a-9efe-5ed830df2640 req-73e2c303-fcf0-40e6-b306-db1b35a1807d service nova] Lock "3d1518d3-bed2-4b7b-af1d-d7da49e92874-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.388855] env[63202]: DEBUG nova.compute.manager [req-7740c387-a001-4c3a-9efe-5ed830df2640 req-73e2c303-fcf0-40e6-b306-db1b35a1807d service nova] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] No waiting events found dispatching network-vif-plugged-06e55a71-f982-49f8-819f-3408e19025de {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 901.388938] env[63202]: WARNING nova.compute.manager [req-7740c387-a001-4c3a-9efe-5ed830df2640 req-73e2c303-fcf0-40e6-b306-db1b35a1807d service nova] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Received unexpected event network-vif-plugged-06e55a71-f982-49f8-819f-3408e19025de for instance with vm_state building and task_state spawning. [ 901.473261] env[63202]: DEBUG nova.network.neutron [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Successfully updated port: 06e55a71-f982-49f8-819f-3408e19025de {{(pid=63202) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 901.583617] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385439, 'name': CreateVM_Task} progress is 99%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.830018] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e73d70cd-9281-4a8e-bc1f-3b8e68b88609 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.836307] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-261bc27a-b95a-45a4-a031-94e51df78ac0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.873312] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd5cbfa7-c2b5-4cf9-ab25-acd221f3381a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.882174] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f35d4fa9-1cbf-428d-86ed-cc054a2847ef {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.898066] env[63202]: DEBUG nova.compute.provider_tree [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 901.979671] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Acquiring lock "refresh_cache-3d1518d3-bed2-4b7b-af1d-d7da49e92874" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.979671] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Acquired lock "refresh_cache-3d1518d3-bed2-4b7b-af1d-d7da49e92874" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.979671] env[63202]: DEBUG nova.network.neutron [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 902.084506] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385439, 'name': CreateVM_Task, 'duration_secs': 0.560904} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.084506] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 36291628-af7d-43cf-8149-09f57df47890] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 902.084506] env[63202]: DEBUG oslo_concurrency.lockutils [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.084506] env[63202]: DEBUG oslo_concurrency.lockutils [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.084506] env[63202]: DEBUG oslo_concurrency.lockutils [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 902.084506] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-59f27d4d-2796-4404-8be7-be84ad95b4cb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.090017] env[63202]: DEBUG oslo_vmware.api [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Waiting for the task: (returnval){ [ 902.090017] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52f992b9-06ce-6471-da80-44533b95007d" [ 902.090017] env[63202]: _type = "Task" [ 902.090017] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.097087] env[63202]: DEBUG oslo_vmware.api [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52f992b9-06ce-6471-da80-44533b95007d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.191542] env[63202]: DEBUG nova.network.neutron [req-c212a235-4621-4050-9c13-2533dd9298fd req-1c07f22b-c89e-40f1-aea7-f7e8f58dfe7b service nova] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Updated VIF entry in instance network info cache for port a0175615-5853-4224-8504-2a296446e39b. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 902.191542] env[63202]: DEBUG nova.network.neutron [req-c212a235-4621-4050-9c13-2533dd9298fd req-1c07f22b-c89e-40f1-aea7-f7e8f58dfe7b service nova] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Updating instance_info_cache with network_info: [{"id": "a0175615-5853-4224-8504-2a296446e39b", "address": "fa:16:3e:d8:ab:76", "network": {"id": "334c5b4e-9931-457d-b98c-1df0cf4ccfa9", "bridge": "br-int", "label": "tempest-ServersTestJSON-720696570-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.136", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1a8d47e87d64f83b4b55d9704531021", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "399f3826-705c-45f7-9fe0-3a08a945151a", "external-id": "nsx-vlan-transportzone-936", "segmentation_id": 936, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0175615-58", "ovs_interfaceid": "a0175615-5853-4224-8504-2a296446e39b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.404539] env[63202]: DEBUG nova.scheduler.client.report [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 902.551228] env[63202]: DEBUG nova.network.neutron [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 902.603044] env[63202]: DEBUG oslo_vmware.api [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52f992b9-06ce-6471-da80-44533b95007d, 'name': SearchDatastore_Task, 'duration_secs': 0.015209} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.603044] env[63202]: DEBUG oslo_concurrency.lockutils [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.603044] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 902.603044] env[63202]: DEBUG oslo_concurrency.lockutils [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.603247] env[63202]: DEBUG oslo_concurrency.lockutils [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.603247] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 902.603247] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-23ab1835-0df2-4a90-a1b9-263f925e260e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.616017] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 902.616017] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 902.616017] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3257c94e-6857-49dd-8702-06a1e5e80595 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.625021] env[63202]: DEBUG oslo_vmware.api [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Waiting for the task: (returnval){ [ 902.625021] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]5215465f-5ae6-310e-c508-7faab46a78b0" [ 902.625021] env[63202]: _type = "Task" [ 902.625021] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.633695] env[63202]: DEBUG oslo_vmware.api [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]5215465f-5ae6-310e-c508-7faab46a78b0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.695307] env[63202]: DEBUG oslo_concurrency.lockutils [req-c212a235-4621-4050-9c13-2533dd9298fd req-1c07f22b-c89e-40f1-aea7-f7e8f58dfe7b service nova] Releasing lock "refresh_cache-2243bb69-0dc5-49cd-b94e-73e703cbadc0" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.695307] env[63202]: DEBUG nova.compute.manager [req-c212a235-4621-4050-9c13-2533dd9298fd req-1c07f22b-c89e-40f1-aea7-f7e8f58dfe7b service nova] [instance: 36291628-af7d-43cf-8149-09f57df47890] Received event network-changed-355608b4-3a2c-4f49-8b9a-6b379a97934d {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 902.695307] env[63202]: DEBUG nova.compute.manager [req-c212a235-4621-4050-9c13-2533dd9298fd req-1c07f22b-c89e-40f1-aea7-f7e8f58dfe7b service nova] [instance: 36291628-af7d-43cf-8149-09f57df47890] Refreshing instance network info cache due to event network-changed-355608b4-3a2c-4f49-8b9a-6b379a97934d. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 902.695307] env[63202]: DEBUG oslo_concurrency.lockutils [req-c212a235-4621-4050-9c13-2533dd9298fd req-1c07f22b-c89e-40f1-aea7-f7e8f58dfe7b service nova] Acquiring lock "refresh_cache-36291628-af7d-43cf-8149-09f57df47890" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.695307] env[63202]: DEBUG oslo_concurrency.lockutils [req-c212a235-4621-4050-9c13-2533dd9298fd req-1c07f22b-c89e-40f1-aea7-f7e8f58dfe7b service nova] Acquired lock "refresh_cache-36291628-af7d-43cf-8149-09f57df47890" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.695560] env[63202]: DEBUG nova.network.neutron [req-c212a235-4621-4050-9c13-2533dd9298fd req-1c07f22b-c89e-40f1-aea7-f7e8f58dfe7b service nova] [instance: 36291628-af7d-43cf-8149-09f57df47890] Refreshing network info cache for port 355608b4-3a2c-4f49-8b9a-6b379a97934d {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 902.907556] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.373s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.908456] env[63202]: DEBUG nova.compute.manager [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 902.913022] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.585s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.913022] env[63202]: INFO nova.compute.claims [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 903.006370] env[63202]: DEBUG nova.network.neutron [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Updating instance_info_cache with network_info: [{"id": "06e55a71-f982-49f8-819f-3408e19025de", "address": "fa:16:3e:87:b4:3f", "network": {"id": "d83c3dc7-a0c9-4401-8b08-99d493f23a8b", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-482332878-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "98fbc2f9559148b8b0e2a2136152f13c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap06e55a71-f9", "ovs_interfaceid": "06e55a71-f982-49f8-819f-3408e19025de", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.138849] env[63202]: DEBUG oslo_vmware.api [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]5215465f-5ae6-310e-c508-7faab46a78b0, 'name': SearchDatastore_Task, 'duration_secs': 0.011738} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.141162] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a12a013-e46f-40ae-b8a3-c09922f702d6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.148191] env[63202]: DEBUG oslo_vmware.api [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Waiting for the task: (returnval){ [ 903.148191] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]524ec434-1bd7-f435-04c4-52defdd6fe5f" [ 903.148191] env[63202]: _type = "Task" [ 903.148191] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.164206] env[63202]: DEBUG oslo_vmware.api [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]524ec434-1bd7-f435-04c4-52defdd6fe5f, 'name': SearchDatastore_Task, 'duration_secs': 0.012924} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.164206] env[63202]: DEBUG oslo_concurrency.lockutils [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.164206] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore2] 36291628-af7d-43cf-8149-09f57df47890/36291628-af7d-43cf-8149-09f57df47890.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 903.164206] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-40d888a0-6b33-4888-8b9e-ba6f48c7271a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.177414] env[63202]: DEBUG oslo_vmware.api [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Waiting for the task: (returnval){ [ 903.177414] env[63202]: value = "task-1385440" [ 903.177414] env[63202]: _type = "Task" [ 903.177414] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.192184] env[63202]: DEBUG oslo_vmware.api [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385440, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.419442] env[63202]: DEBUG nova.compute.utils [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 903.422926] env[63202]: DEBUG nova.compute.manager [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 903.423277] env[63202]: DEBUG nova.network.neutron [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 903.427137] env[63202]: DEBUG nova.compute.manager [req-d0e37034-4f7b-40e5-8d00-8972112cd9de req-4d350ed2-e46c-49b0-9ec0-38417427b660 service nova] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Received event network-changed-06e55a71-f982-49f8-819f-3408e19025de {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 903.427137] env[63202]: DEBUG nova.compute.manager [req-d0e37034-4f7b-40e5-8d00-8972112cd9de req-4d350ed2-e46c-49b0-9ec0-38417427b660 service nova] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Refreshing instance network info cache due to event network-changed-06e55a71-f982-49f8-819f-3408e19025de. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 903.427137] env[63202]: DEBUG oslo_concurrency.lockutils [req-d0e37034-4f7b-40e5-8d00-8972112cd9de req-4d350ed2-e46c-49b0-9ec0-38417427b660 service nova] Acquiring lock "refresh_cache-3d1518d3-bed2-4b7b-af1d-d7da49e92874" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.446480] env[63202]: DEBUG nova.network.neutron [req-c212a235-4621-4050-9c13-2533dd9298fd req-1c07f22b-c89e-40f1-aea7-f7e8f58dfe7b service nova] [instance: 36291628-af7d-43cf-8149-09f57df47890] Updated VIF entry in instance network info cache for port 355608b4-3a2c-4f49-8b9a-6b379a97934d. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 903.446480] env[63202]: DEBUG nova.network.neutron [req-c212a235-4621-4050-9c13-2533dd9298fd req-1c07f22b-c89e-40f1-aea7-f7e8f58dfe7b service nova] [instance: 36291628-af7d-43cf-8149-09f57df47890] Updating instance_info_cache with network_info: [{"id": "355608b4-3a2c-4f49-8b9a-6b379a97934d", "address": "fa:16:3e:59:3b:cb", "network": {"id": "d83c3dc7-a0c9-4401-8b08-99d493f23a8b", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-482332878-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "98fbc2f9559148b8b0e2a2136152f13c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap355608b4-3a", "ovs_interfaceid": "355608b4-3a2c-4f49-8b9a-6b379a97934d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.511024] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Releasing lock "refresh_cache-3d1518d3-bed2-4b7b-af1d-d7da49e92874" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.511024] env[63202]: DEBUG nova.compute.manager [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Instance network_info: |[{"id": "06e55a71-f982-49f8-819f-3408e19025de", "address": "fa:16:3e:87:b4:3f", "network": {"id": "d83c3dc7-a0c9-4401-8b08-99d493f23a8b", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-482332878-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "98fbc2f9559148b8b0e2a2136152f13c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap06e55a71-f9", "ovs_interfaceid": "06e55a71-f982-49f8-819f-3408e19025de", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 903.511287] env[63202]: DEBUG oslo_concurrency.lockutils [req-d0e37034-4f7b-40e5-8d00-8972112cd9de req-4d350ed2-e46c-49b0-9ec0-38417427b660 service nova] Acquired lock "refresh_cache-3d1518d3-bed2-4b7b-af1d-d7da49e92874" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.511287] env[63202]: DEBUG nova.network.neutron [req-d0e37034-4f7b-40e5-8d00-8972112cd9de req-4d350ed2-e46c-49b0-9ec0-38417427b660 service nova] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Refreshing network info cache for port 06e55a71-f982-49f8-819f-3408e19025de {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 903.511287] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:87:b4:3f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f3ccbdbb-8b49-4a26-913f-2a448b72280f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '06e55a71-f982-49f8-819f-3408e19025de', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 903.519243] env[63202]: DEBUG oslo.service.loopingcall [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 903.520541] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 903.520892] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-346eb841-294b-4d6b-8bf2-dca36490c9ee {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.547024] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 903.547024] env[63202]: value = "task-1385441" [ 903.547024] env[63202]: _type = "Task" [ 903.547024] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.556102] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385441, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.691918] env[63202]: DEBUG oslo_vmware.api [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385440, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.798599] env[63202]: DEBUG nova.policy [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'faabe4f6570d41598fd1f2b2f0d6df11', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '98fbc2f9559148b8b0e2a2136152f13c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 903.923766] env[63202]: DEBUG nova.compute.manager [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 903.949736] env[63202]: DEBUG oslo_concurrency.lockutils [req-c212a235-4621-4050-9c13-2533dd9298fd req-1c07f22b-c89e-40f1-aea7-f7e8f58dfe7b service nova] Releasing lock "refresh_cache-36291628-af7d-43cf-8149-09f57df47890" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.061280] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385441, 'name': CreateVM_Task, 'duration_secs': 0.433546} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.061448] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 904.062179] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.062305] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.062563] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 904.063376] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9d46512a-b454-4bec-8b48-819a1472f4c5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.072060] env[63202]: DEBUG oslo_vmware.api [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Waiting for the task: (returnval){ [ 904.072060] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52f4826f-13ff-8c13-72c5-1faf71523659" [ 904.072060] env[63202]: _type = "Task" [ 904.072060] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.086673] env[63202]: DEBUG oslo_vmware.api [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52f4826f-13ff-8c13-72c5-1faf71523659, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.200359] env[63202]: DEBUG oslo_vmware.api [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385440, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.562152} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.200720] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore2] 36291628-af7d-43cf-8149-09f57df47890/36291628-af7d-43cf-8149-09f57df47890.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 904.200855] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 904.202099] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9b811ba3-3a49-40a7-ae3f-3ece654a240b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.216067] env[63202]: DEBUG oslo_vmware.api [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Waiting for the task: (returnval){ [ 904.216067] env[63202]: value = "task-1385442" [ 904.216067] env[63202]: _type = "Task" [ 904.216067] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.226310] env[63202]: DEBUG oslo_vmware.api [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385442, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.263386] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faae8a62-fd28-4083-b2dd-806f51a0389c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.279090] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3283a74e-1207-413f-a103-a6fa9d37d55d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.309040] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95e438dc-ba91-441a-95b6-83d3f8f26ba4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.323209] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-683ea4b5-ab4d-4a2f-9fe7-4524853ac982 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.330719] env[63202]: DEBUG nova.network.neutron [req-d0e37034-4f7b-40e5-8d00-8972112cd9de req-4d350ed2-e46c-49b0-9ec0-38417427b660 service nova] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Updated VIF entry in instance network info cache for port 06e55a71-f982-49f8-819f-3408e19025de. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 904.331143] env[63202]: DEBUG nova.network.neutron [req-d0e37034-4f7b-40e5-8d00-8972112cd9de req-4d350ed2-e46c-49b0-9ec0-38417427b660 service nova] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Updating instance_info_cache with network_info: [{"id": "06e55a71-f982-49f8-819f-3408e19025de", "address": "fa:16:3e:87:b4:3f", "network": {"id": "d83c3dc7-a0c9-4401-8b08-99d493f23a8b", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-482332878-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "98fbc2f9559148b8b0e2a2136152f13c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap06e55a71-f9", "ovs_interfaceid": "06e55a71-f982-49f8-819f-3408e19025de", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.344894] env[63202]: DEBUG nova.compute.provider_tree [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 904.351075] env[63202]: DEBUG oslo_concurrency.lockutils [req-d0e37034-4f7b-40e5-8d00-8972112cd9de req-4d350ed2-e46c-49b0-9ec0-38417427b660 service nova] Releasing lock "refresh_cache-3d1518d3-bed2-4b7b-af1d-d7da49e92874" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.453270] env[63202]: DEBUG nova.network.neutron [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Successfully created port: 38e58a7b-1f81-49a0-9b7d-25c507dc32e7 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 904.585221] env[63202]: DEBUG oslo_vmware.api [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52f4826f-13ff-8c13-72c5-1faf71523659, 'name': SearchDatastore_Task, 'duration_secs': 0.014898} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.585554] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.585783] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 904.586048] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.586201] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.586375] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 904.586649] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3b47dab0-24c3-4637-acf2-ed496a1bcf7d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.603375] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 904.603567] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 904.604339] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3bd20239-5de2-4451-a308-6cada0eb47f1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.611986] env[63202]: DEBUG oslo_vmware.api [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Waiting for the task: (returnval){ [ 904.611986] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52a72538-7323-4daf-9a0f-48cadc0ff294" [ 904.611986] env[63202]: _type = "Task" [ 904.611986] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.623066] env[63202]: DEBUG oslo_vmware.api [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52a72538-7323-4daf-9a0f-48cadc0ff294, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.727421] env[63202]: DEBUG oslo_vmware.api [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385442, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073961} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.727421] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 904.728157] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e3d232a-5925-41b7-ad39-002e7b1195b0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.755329] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] Reconfiguring VM instance instance-00000044 to attach disk [datastore2] 36291628-af7d-43cf-8149-09f57df47890/36291628-af7d-43cf-8149-09f57df47890.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 904.755329] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-89c145c8-6ee4-4280-8d98-2a9e8aaadef5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.776499] env[63202]: DEBUG oslo_vmware.api [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Waiting for the task: (returnval){ [ 904.776499] env[63202]: value = "task-1385443" [ 904.776499] env[63202]: _type = "Task" [ 904.776499] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.785759] env[63202]: DEBUG oslo_vmware.api [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385443, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.854017] env[63202]: DEBUG nova.scheduler.client.report [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 904.934100] env[63202]: DEBUG nova.compute.manager [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 904.982227] env[63202]: DEBUG nova.virt.hardware [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 904.982227] env[63202]: DEBUG nova.virt.hardware [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 904.982227] env[63202]: DEBUG nova.virt.hardware [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 904.982717] env[63202]: DEBUG nova.virt.hardware [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 904.982856] env[63202]: DEBUG nova.virt.hardware [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 904.983278] env[63202]: DEBUG nova.virt.hardware [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 904.983669] env[63202]: DEBUG nova.virt.hardware [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 904.983920] env[63202]: DEBUG nova.virt.hardware [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 904.984172] env[63202]: DEBUG nova.virt.hardware [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 904.984404] env[63202]: DEBUG nova.virt.hardware [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 904.984631] env[63202]: DEBUG nova.virt.hardware [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 904.985921] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd622ebd-0931-4e3c-a927-e7b236e9a993 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.003645] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-409586c3-518f-439c-8600-2afc403e53bc {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.122145] env[63202]: DEBUG oslo_vmware.api [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52a72538-7323-4daf-9a0f-48cadc0ff294, 'name': SearchDatastore_Task, 'duration_secs': 0.02644} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.122941] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-206b0fb9-b3c0-4771-bebe-95cb32aa423d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.128641] env[63202]: DEBUG oslo_vmware.api [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Waiting for the task: (returnval){ [ 905.128641] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]5203f169-8cae-eb95-aeb7-b472eb1320d6" [ 905.128641] env[63202]: _type = "Task" [ 905.128641] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.137220] env[63202]: DEBUG oslo_vmware.api [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]5203f169-8cae-eb95-aeb7-b472eb1320d6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.292666] env[63202]: DEBUG oslo_vmware.api [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385443, 'name': ReconfigVM_Task, 'duration_secs': 0.302791} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.292666] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] Reconfigured VM instance instance-00000044 to attach disk [datastore2] 36291628-af7d-43cf-8149-09f57df47890/36291628-af7d-43cf-8149-09f57df47890.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 905.292666] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-279e01ad-2334-4fbd-b6b9-96360854efe2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.302431] env[63202]: DEBUG oslo_vmware.api [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Waiting for the task: (returnval){ [ 905.302431] env[63202]: value = "task-1385444" [ 905.302431] env[63202]: _type = "Task" [ 905.302431] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.316520] env[63202]: DEBUG oslo_vmware.api [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385444, 'name': Rename_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.358469] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.447s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.360839] env[63202]: DEBUG nova.compute.manager [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 905.362992] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.617s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.369089] env[63202]: INFO nova.compute.claims [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 905.640186] env[63202]: DEBUG oslo_vmware.api [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]5203f169-8cae-eb95-aeb7-b472eb1320d6, 'name': SearchDatastore_Task, 'duration_secs': 0.041541} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.640510] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.641249] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore2] 3d1518d3-bed2-4b7b-af1d-d7da49e92874/3d1518d3-bed2-4b7b-af1d-d7da49e92874.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 905.641249] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5adcee25-2b4a-486a-ae72-1be644811655 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.650063] env[63202]: DEBUG oslo_vmware.api [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Waiting for the task: (returnval){ [ 905.650063] env[63202]: value = "task-1385445" [ 905.650063] env[63202]: _type = "Task" [ 905.650063] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.659276] env[63202]: DEBUG oslo_vmware.api [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385445, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.814918] env[63202]: DEBUG oslo_vmware.api [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385444, 'name': Rename_Task} progress is 99%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.872661] env[63202]: DEBUG nova.compute.utils [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 905.877247] env[63202]: DEBUG nova.compute.manager [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 905.877247] env[63202]: DEBUG nova.network.neutron [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 905.956916] env[63202]: DEBUG nova.policy [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2d292997e4c34051acc645b4a65dc369', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5e914f87450949b1a39866e8cfa3a0eb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 906.165112] env[63202]: DEBUG oslo_vmware.api [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385445, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.265976] env[63202]: DEBUG nova.compute.manager [req-6d6d77f8-c950-4aa2-9ead-2e5e2cbc23ad req-f77b39af-426b-424d-bec9-926175f24a41 service nova] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Received event network-vif-plugged-38e58a7b-1f81-49a0-9b7d-25c507dc32e7 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 906.266136] env[63202]: DEBUG oslo_concurrency.lockutils [req-6d6d77f8-c950-4aa2-9ead-2e5e2cbc23ad req-f77b39af-426b-424d-bec9-926175f24a41 service nova] Acquiring lock "61a6528e-4e4e-49b7-be7b-dd30bd8023d8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.266318] env[63202]: DEBUG oslo_concurrency.lockutils [req-6d6d77f8-c950-4aa2-9ead-2e5e2cbc23ad req-f77b39af-426b-424d-bec9-926175f24a41 service nova] Lock "61a6528e-4e4e-49b7-be7b-dd30bd8023d8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.266502] env[63202]: DEBUG oslo_concurrency.lockutils [req-6d6d77f8-c950-4aa2-9ead-2e5e2cbc23ad req-f77b39af-426b-424d-bec9-926175f24a41 service nova] Lock "61a6528e-4e4e-49b7-be7b-dd30bd8023d8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.266638] env[63202]: DEBUG nova.compute.manager [req-6d6d77f8-c950-4aa2-9ead-2e5e2cbc23ad req-f77b39af-426b-424d-bec9-926175f24a41 service nova] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] No waiting events found dispatching network-vif-plugged-38e58a7b-1f81-49a0-9b7d-25c507dc32e7 {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 906.266798] env[63202]: WARNING nova.compute.manager [req-6d6d77f8-c950-4aa2-9ead-2e5e2cbc23ad req-f77b39af-426b-424d-bec9-926175f24a41 service nova] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Received unexpected event network-vif-plugged-38e58a7b-1f81-49a0-9b7d-25c507dc32e7 for instance with vm_state building and task_state spawning. [ 906.324771] env[63202]: DEBUG oslo_vmware.api [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385444, 'name': Rename_Task} progress is 99%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.379932] env[63202]: DEBUG nova.compute.manager [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 906.425312] env[63202]: DEBUG nova.network.neutron [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Successfully updated port: 38e58a7b-1f81-49a0-9b7d-25c507dc32e7 {{(pid=63202) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 906.430952] env[63202]: DEBUG nova.network.neutron [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Successfully created port: aa2cbb5d-3724-4754-922b-da0798168b84 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 906.668534] env[63202]: DEBUG oslo_vmware.api [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385445, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.562626} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.669542] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore2] 3d1518d3-bed2-4b7b-af1d-d7da49e92874/3d1518d3-bed2-4b7b-af1d-d7da49e92874.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 906.669542] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 906.669754] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9af3db2c-c449-4cb3-94f7-2d745f8e57d7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.683429] env[63202]: DEBUG oslo_vmware.api [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Waiting for the task: (returnval){ [ 906.683429] env[63202]: value = "task-1385446" [ 906.683429] env[63202]: _type = "Task" [ 906.683429] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.697730] env[63202]: DEBUG oslo_vmware.api [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385446, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.710182] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b76f1813-d68e-4781-8d39-a493d376b984 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.719103] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb97d019-dac7-4403-9b3f-2e5bdb20ed26 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.754994] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57609cd1-321d-4479-951c-4d16c05ffe4e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.765039] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cc28ca9-4f74-4703-93f2-db87a9267fcd {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.780749] env[63202]: DEBUG nova.compute.provider_tree [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 906.818661] env[63202]: DEBUG oslo_vmware.api [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385444, 'name': Rename_Task, 'duration_secs': 1.166067} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.818661] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 906.818906] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7efcf7a4-a0f2-4068-b6ee-14ed85cf3cbd {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.829523] env[63202]: DEBUG oslo_vmware.api [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Waiting for the task: (returnval){ [ 906.829523] env[63202]: value = "task-1385447" [ 906.829523] env[63202]: _type = "Task" [ 906.829523] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.839901] env[63202]: DEBUG oslo_vmware.api [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385447, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.938638] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Acquiring lock "refresh_cache-61a6528e-4e4e-49b7-be7b-dd30bd8023d8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.938638] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Acquired lock "refresh_cache-61a6528e-4e4e-49b7-be7b-dd30bd8023d8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.938638] env[63202]: DEBUG nova.network.neutron [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 907.197159] env[63202]: DEBUG oslo_vmware.api [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385446, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074274} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.198077] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 907.200718] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00860b29-c6cf-44d9-8ea6-5b1bfc5737c7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.235101] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Reconfiguring VM instance instance-00000045 to attach disk [datastore2] 3d1518d3-bed2-4b7b-af1d-d7da49e92874/3d1518d3-bed2-4b7b-af1d-d7da49e92874.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 907.236708] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2240fdab-9033-47c5-9531-b6c074f70d89 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.260835] env[63202]: DEBUG oslo_vmware.api [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Waiting for the task: (returnval){ [ 907.260835] env[63202]: value = "task-1385448" [ 907.260835] env[63202]: _type = "Task" [ 907.260835] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.273166] env[63202]: DEBUG oslo_vmware.api [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385448, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.286287] env[63202]: DEBUG nova.scheduler.client.report [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 907.342900] env[63202]: DEBUG oslo_vmware.api [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385447, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.404240] env[63202]: DEBUG nova.compute.manager [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 907.436479] env[63202]: DEBUG nova.virt.hardware [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 907.436836] env[63202]: DEBUG nova.virt.hardware [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 907.437166] env[63202]: DEBUG nova.virt.hardware [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 907.437407] env[63202]: DEBUG nova.virt.hardware [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 907.437613] env[63202]: DEBUG nova.virt.hardware [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 907.437774] env[63202]: DEBUG nova.virt.hardware [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 907.438013] env[63202]: DEBUG nova.virt.hardware [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 907.438186] env[63202]: DEBUG nova.virt.hardware [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 907.438353] env[63202]: DEBUG nova.virt.hardware [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 907.438512] env[63202]: DEBUG nova.virt.hardware [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 907.438682] env[63202]: DEBUG nova.virt.hardware [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 907.439814] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7431aa38-399e-4f27-940e-c0a532e6720e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.457016] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dbaa8d5-05e3-4d04-855a-2f8f1d6cf475 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.505198] env[63202]: DEBUG nova.network.neutron [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 907.512755] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Acquiring lock "f3128c09-3680-4b0b-b463-3d6cd203fcf4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.513060] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Lock "f3128c09-3680-4b0b-b463-3d6cd203fcf4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.772258] env[63202]: DEBUG oslo_vmware.api [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385448, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.779737] env[63202]: DEBUG nova.network.neutron [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Updating instance_info_cache with network_info: [{"id": "38e58a7b-1f81-49a0-9b7d-25c507dc32e7", "address": "fa:16:3e:b8:fc:3e", "network": {"id": "d83c3dc7-a0c9-4401-8b08-99d493f23a8b", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-482332878-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "98fbc2f9559148b8b0e2a2136152f13c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38e58a7b-1f", "ovs_interfaceid": "38e58a7b-1f81-49a0-9b7d-25c507dc32e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.790228] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.427s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.790810] env[63202]: DEBUG nova.compute.manager [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 907.794102] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 14.002s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.849898] env[63202]: DEBUG oslo_vmware.api [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385447, 'name': PowerOnVM_Task, 'duration_secs': 0.638786} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.850383] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 907.850599] env[63202]: INFO nova.compute.manager [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] Took 9.64 seconds to spawn the instance on the hypervisor. [ 907.850726] env[63202]: DEBUG nova.compute.manager [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 907.851550] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-637237b1-e6d1-40ff-995e-54ed1bf36249 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.098343] env[63202]: DEBUG nova.compute.manager [req-c0dc7a9a-f3c7-465f-a991-daecbd5a29ca req-3a2ff616-e7ab-47b5-ba4c-09438cc2ff08 service nova] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Received event network-vif-plugged-aa2cbb5d-3724-4754-922b-da0798168b84 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 908.098536] env[63202]: DEBUG oslo_concurrency.lockutils [req-c0dc7a9a-f3c7-465f-a991-daecbd5a29ca req-3a2ff616-e7ab-47b5-ba4c-09438cc2ff08 service nova] Acquiring lock "7a72a1ab-3b3d-450d-a3a6-572b0035eb0e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.098749] env[63202]: DEBUG oslo_concurrency.lockutils [req-c0dc7a9a-f3c7-465f-a991-daecbd5a29ca req-3a2ff616-e7ab-47b5-ba4c-09438cc2ff08 service nova] Lock "7a72a1ab-3b3d-450d-a3a6-572b0035eb0e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.098918] env[63202]: DEBUG oslo_concurrency.lockutils [req-c0dc7a9a-f3c7-465f-a991-daecbd5a29ca req-3a2ff616-e7ab-47b5-ba4c-09438cc2ff08 service nova] Lock "7a72a1ab-3b3d-450d-a3a6-572b0035eb0e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.100534] env[63202]: DEBUG nova.compute.manager [req-c0dc7a9a-f3c7-465f-a991-daecbd5a29ca req-3a2ff616-e7ab-47b5-ba4c-09438cc2ff08 service nova] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] No waiting events found dispatching network-vif-plugged-aa2cbb5d-3724-4754-922b-da0798168b84 {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 908.100930] env[63202]: WARNING nova.compute.manager [req-c0dc7a9a-f3c7-465f-a991-daecbd5a29ca req-3a2ff616-e7ab-47b5-ba4c-09438cc2ff08 service nova] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Received unexpected event network-vif-plugged-aa2cbb5d-3724-4754-922b-da0798168b84 for instance with vm_state building and task_state spawning. [ 908.184400] env[63202]: DEBUG nova.network.neutron [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Successfully updated port: aa2cbb5d-3724-4754-922b-da0798168b84 {{(pid=63202) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 908.272036] env[63202]: DEBUG oslo_vmware.api [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385448, 'name': ReconfigVM_Task, 'duration_secs': 0.778044} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.272422] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Reconfigured VM instance instance-00000045 to attach disk [datastore2] 3d1518d3-bed2-4b7b-af1d-d7da49e92874/3d1518d3-bed2-4b7b-af1d-d7da49e92874.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 908.273135] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1d01662f-c4e9-4dad-a72d-2a430531d80f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.280640] env[63202]: DEBUG oslo_vmware.api [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Waiting for the task: (returnval){ [ 908.280640] env[63202]: value = "task-1385449" [ 908.280640] env[63202]: _type = "Task" [ 908.280640] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.284267] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Releasing lock "refresh_cache-61a6528e-4e4e-49b7-be7b-dd30bd8023d8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.284633] env[63202]: DEBUG nova.compute.manager [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Instance network_info: |[{"id": "38e58a7b-1f81-49a0-9b7d-25c507dc32e7", "address": "fa:16:3e:b8:fc:3e", "network": {"id": "d83c3dc7-a0c9-4401-8b08-99d493f23a8b", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-482332878-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "98fbc2f9559148b8b0e2a2136152f13c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38e58a7b-1f", "ovs_interfaceid": "38e58a7b-1f81-49a0-9b7d-25c507dc32e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 908.285182] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b8:fc:3e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f3ccbdbb-8b49-4a26-913f-2a448b72280f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '38e58a7b-1f81-49a0-9b7d-25c507dc32e7', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 908.297259] env[63202]: DEBUG oslo.service.loopingcall [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 908.298986] env[63202]: DEBUG nova.compute.utils [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 908.300383] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 908.306240] env[63202]: DEBUG nova.compute.manager [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 908.306418] env[63202]: DEBUG nova.network.neutron [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 908.313698] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f73a278c-d9ad-453d-b893-28109f675917 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.334507] env[63202]: DEBUG oslo_vmware.api [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385449, 'name': Rename_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.335666] env[63202]: DEBUG nova.compute.manager [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 908.345374] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 908.345374] env[63202]: value = "task-1385450" [ 908.345374] env[63202]: _type = "Task" [ 908.345374] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.355484] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385450, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.372986] env[63202]: INFO nova.compute.manager [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] Took 20.94 seconds to build instance. [ 908.429860] env[63202]: DEBUG nova.policy [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1236175893a3401cb01b04604168f1e0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '22352c25bca8416a948014391a5389ac', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 908.454463] env[63202]: DEBUG nova.compute.manager [req-70cf7ec9-98b7-4970-9955-545798679e09 req-2dd98e3e-6a65-42ed-b838-253855d8adb5 service nova] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Received event network-changed-38e58a7b-1f81-49a0-9b7d-25c507dc32e7 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 908.454877] env[63202]: DEBUG nova.compute.manager [req-70cf7ec9-98b7-4970-9955-545798679e09 req-2dd98e3e-6a65-42ed-b838-253855d8adb5 service nova] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Refreshing instance network info cache due to event network-changed-38e58a7b-1f81-49a0-9b7d-25c507dc32e7. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 908.454877] env[63202]: DEBUG oslo_concurrency.lockutils [req-70cf7ec9-98b7-4970-9955-545798679e09 req-2dd98e3e-6a65-42ed-b838-253855d8adb5 service nova] Acquiring lock "refresh_cache-61a6528e-4e4e-49b7-be7b-dd30bd8023d8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 908.455298] env[63202]: DEBUG oslo_concurrency.lockutils [req-70cf7ec9-98b7-4970-9955-545798679e09 req-2dd98e3e-6a65-42ed-b838-253855d8adb5 service nova] Acquired lock "refresh_cache-61a6528e-4e4e-49b7-be7b-dd30bd8023d8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.455439] env[63202]: DEBUG nova.network.neutron [req-70cf7ec9-98b7-4970-9955-545798679e09 req-2dd98e3e-6a65-42ed-b838-253855d8adb5 service nova] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Refreshing network info cache for port 38e58a7b-1f81-49a0-9b7d-25c507dc32e7 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 908.691058] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquiring lock "refresh_cache-7a72a1ab-3b3d-450d-a3a6-572b0035eb0e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 908.691183] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquired lock "refresh_cache-7a72a1ab-3b3d-450d-a3a6-572b0035eb0e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.691880] env[63202]: DEBUG nova.network.neutron [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 908.756286] env[63202]: DEBUG nova.network.neutron [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Successfully created port: c7da6929-1bd1-4e2b-bcb8-609863cf6861 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 908.796252] env[63202]: DEBUG oslo_vmware.api [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385449, 'name': Rename_Task, 'duration_secs': 0.21822} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.796630] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 908.797127] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-58b652ee-b64b-4901-89af-eb256b6c8c8c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.806511] env[63202]: DEBUG oslo_vmware.api [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Waiting for the task: (returnval){ [ 908.806511] env[63202]: value = "task-1385451" [ 908.806511] env[63202]: _type = "Task" [ 908.806511] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.816542] env[63202]: DEBUG oslo_vmware.api [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385451, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.857009] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385450, 'name': CreateVM_Task, 'duration_secs': 0.36223} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.857450] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 908.858226] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 908.858474] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.858888] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 908.859536] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9be46d88-379c-4d49-bb78-7e5b9d5ea37c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.863274] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 908.863434] env[63202]: WARNING nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance a800fa5d-46a1-4a96-8ea2-26e34b869aa6 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 908.863565] env[63202]: WARNING nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 908.863691] env[63202]: WARNING nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 50854a4e-8e4e-431b-8220-01bf8906b1c0 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 908.863807] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 2243bb69-0dc5-49cd-b94e-73e703cbadc0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 908.863920] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 36291628-af7d-43cf-8149-09f57df47890 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 908.868120] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 3d1518d3-bed2-4b7b-af1d-d7da49e92874 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 908.868288] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 61a6528e-4e4e-49b7-be7b-dd30bd8023d8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 908.868437] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 908.868559] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance e775e5e1-521a-4fc7-80e6-bcb6a70516c5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 908.873997] env[63202]: DEBUG oslo_vmware.api [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Waiting for the task: (returnval){ [ 908.873997] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]521f67ab-2883-740c-01e7-d3f05e2314f0" [ 908.873997] env[63202]: _type = "Task" [ 908.873997] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.879700] env[63202]: DEBUG oslo_concurrency.lockutils [None req-86892c1d-d929-40fb-9caf-d1611bf524c3 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Lock "36291628-af7d-43cf-8149-09f57df47890" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 65.546s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.895844] env[63202]: DEBUG oslo_vmware.api [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]521f67ab-2883-740c-01e7-d3f05e2314f0, 'name': SearchDatastore_Task, 'duration_secs': 0.013089} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.896177] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.896423] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 908.896674] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 908.896838] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.897027] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 908.897596] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d939f2d4-3809-4494-a08b-52803966eafc {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.907727] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 908.907859] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 908.908618] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d27346b-090f-4d05-81f9-7f9dfc23d735 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.915803] env[63202]: DEBUG oslo_vmware.api [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Waiting for the task: (returnval){ [ 908.915803] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]524a537b-e469-643f-ee01-e4ca71389e72" [ 908.915803] env[63202]: _type = "Task" [ 908.915803] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.925225] env[63202]: DEBUG oslo_vmware.api [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]524a537b-e469-643f-ee01-e4ca71389e72, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.199092] env[63202]: DEBUG nova.network.neutron [req-70cf7ec9-98b7-4970-9955-545798679e09 req-2dd98e3e-6a65-42ed-b838-253855d8adb5 service nova] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Updated VIF entry in instance network info cache for port 38e58a7b-1f81-49a0-9b7d-25c507dc32e7. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 909.199569] env[63202]: DEBUG nova.network.neutron [req-70cf7ec9-98b7-4970-9955-545798679e09 req-2dd98e3e-6a65-42ed-b838-253855d8adb5 service nova] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Updating instance_info_cache with network_info: [{"id": "38e58a7b-1f81-49a0-9b7d-25c507dc32e7", "address": "fa:16:3e:b8:fc:3e", "network": {"id": "d83c3dc7-a0c9-4401-8b08-99d493f23a8b", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-482332878-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "98fbc2f9559148b8b0e2a2136152f13c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38e58a7b-1f", "ovs_interfaceid": "38e58a7b-1f81-49a0-9b7d-25c507dc32e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 909.237311] env[63202]: DEBUG nova.network.neutron [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 909.317622] env[63202]: DEBUG oslo_vmware.api [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385451, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.345909] env[63202]: DEBUG nova.compute.manager [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 909.371704] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 753d190b-f4a4-4438-bc98-94564e3ec73d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 909.383619] env[63202]: DEBUG nova.virt.hardware [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 909.384084] env[63202]: DEBUG nova.virt.hardware [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 909.384251] env[63202]: DEBUG nova.virt.hardware [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 909.384435] env[63202]: DEBUG nova.virt.hardware [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 909.384576] env[63202]: DEBUG nova.virt.hardware [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 909.384945] env[63202]: DEBUG nova.virt.hardware [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 909.385199] env[63202]: DEBUG nova.virt.hardware [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 909.385362] env[63202]: DEBUG nova.virt.hardware [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 909.385526] env[63202]: DEBUG nova.virt.hardware [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 909.385684] env[63202]: DEBUG nova.virt.hardware [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 909.386061] env[63202]: DEBUG nova.virt.hardware [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 909.387136] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac80e280-3721-41f1-8e99-356a2c23245d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.394020] env[63202]: DEBUG nova.compute.manager [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 909.402723] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1083999-0e10-4729-88c3-46c818805447 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.411534] env[63202]: DEBUG nova.network.neutron [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Updating instance_info_cache with network_info: [{"id": "aa2cbb5d-3724-4754-922b-da0798168b84", "address": "fa:16:3e:6b:99:b5", "network": {"id": "e6c760f6-ded9-4905-b303-dbcde0d037b9", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1433567753-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e914f87450949b1a39866e8cfa3a0eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "26472e27-9835-4f87-ab7f-ca24dfee4e83", "external-id": "nsx-vlan-transportzone-335", "segmentation_id": 335, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa2cbb5d-37", "ovs_interfaceid": "aa2cbb5d-3724-4754-922b-da0798168b84", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 909.438292] env[63202]: DEBUG oslo_vmware.api [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]524a537b-e469-643f-ee01-e4ca71389e72, 'name': SearchDatastore_Task, 'duration_secs': 0.010457} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.438875] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9660cacc-59bd-47f5-bc4f-025f86033593 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.446732] env[63202]: DEBUG oslo_vmware.api [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Waiting for the task: (returnval){ [ 909.446732] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52edebc8-da2d-c851-0703-cef5d4cbd3b3" [ 909.446732] env[63202]: _type = "Task" [ 909.446732] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.456140] env[63202]: DEBUG oslo_vmware.api [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52edebc8-da2d-c851-0703-cef5d4cbd3b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.703602] env[63202]: DEBUG oslo_concurrency.lockutils [req-70cf7ec9-98b7-4970-9955-545798679e09 req-2dd98e3e-6a65-42ed-b838-253855d8adb5 service nova] Releasing lock "refresh_cache-61a6528e-4e4e-49b7-be7b-dd30bd8023d8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.820104] env[63202]: DEBUG oslo_vmware.api [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385451, 'name': PowerOnVM_Task, 'duration_secs': 0.539735} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.820104] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 909.820104] env[63202]: INFO nova.compute.manager [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Took 9.19 seconds to spawn the instance on the hypervisor. [ 909.820817] env[63202]: DEBUG nova.compute.manager [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 909.821097] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0a3937c-88ec-45c7-999c-21b8ae9ad082 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.876129] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance af2bedc2-28ee-4679-ae38-1cceb2af05d6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 909.912546] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.915078] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Releasing lock "refresh_cache-7a72a1ab-3b3d-450d-a3a6-572b0035eb0e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.915362] env[63202]: DEBUG nova.compute.manager [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Instance network_info: |[{"id": "aa2cbb5d-3724-4754-922b-da0798168b84", "address": "fa:16:3e:6b:99:b5", "network": {"id": "e6c760f6-ded9-4905-b303-dbcde0d037b9", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1433567753-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e914f87450949b1a39866e8cfa3a0eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "26472e27-9835-4f87-ab7f-ca24dfee4e83", "external-id": "nsx-vlan-transportzone-335", "segmentation_id": 335, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa2cbb5d-37", "ovs_interfaceid": "aa2cbb5d-3724-4754-922b-da0798168b84", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 909.915969] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6b:99:b5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '26472e27-9835-4f87-ab7f-ca24dfee4e83', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'aa2cbb5d-3724-4754-922b-da0798168b84', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 909.923425] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Creating folder: Project (5e914f87450949b1a39866e8cfa3a0eb). Parent ref: group-v294090. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 909.924028] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9745247e-84c9-43e3-b715-6e5375ca762b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.958258] env[63202]: DEBUG oslo_vmware.api [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52edebc8-da2d-c851-0703-cef5d4cbd3b3, 'name': SearchDatastore_Task, 'duration_secs': 0.011502} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.958340] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.958553] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore2] 61a6528e-4e4e-49b7-be7b-dd30bd8023d8/61a6528e-4e4e-49b7-be7b-dd30bd8023d8.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 909.958812] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-da26f8a5-e75f-445e-9950-40cefe4bafeb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.995108] env[63202]: DEBUG oslo_vmware.api [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Waiting for the task: (returnval){ [ 909.995108] env[63202]: value = "task-1385453" [ 909.995108] env[63202]: _type = "Task" [ 909.995108] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.001168] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Created folder: Project (5e914f87450949b1a39866e8cfa3a0eb) in parent group-v294090. [ 910.001707] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Creating folder: Instances. Parent ref: group-v294135. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 910.002482] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-751d584a-d55d-4c4f-8996-f5f2d10d945b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.007863] env[63202]: DEBUG oslo_vmware.api [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385453, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.013637] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Created folder: Instances in parent group-v294135. [ 910.013925] env[63202]: DEBUG oslo.service.loopingcall [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 910.014174] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 910.014434] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-26e35b25-37b2-41c6-be5f-0546474eec0d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.035644] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 910.035644] env[63202]: value = "task-1385455" [ 910.035644] env[63202]: _type = "Task" [ 910.035644] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.044694] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385455, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.136119] env[63202]: DEBUG nova.compute.manager [req-a6f4a3f5-4a62-42c8-ba2d-34529a0cdc94 req-0260b0b6-7f95-45a5-878f-7daa4b349f90 service nova] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Received event network-changed-aa2cbb5d-3724-4754-922b-da0798168b84 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 910.136119] env[63202]: DEBUG nova.compute.manager [req-a6f4a3f5-4a62-42c8-ba2d-34529a0cdc94 req-0260b0b6-7f95-45a5-878f-7daa4b349f90 service nova] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Refreshing instance network info cache due to event network-changed-aa2cbb5d-3724-4754-922b-da0798168b84. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 910.136204] env[63202]: DEBUG oslo_concurrency.lockutils [req-a6f4a3f5-4a62-42c8-ba2d-34529a0cdc94 req-0260b0b6-7f95-45a5-878f-7daa4b349f90 service nova] Acquiring lock "refresh_cache-7a72a1ab-3b3d-450d-a3a6-572b0035eb0e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.136292] env[63202]: DEBUG oslo_concurrency.lockutils [req-a6f4a3f5-4a62-42c8-ba2d-34529a0cdc94 req-0260b0b6-7f95-45a5-878f-7daa4b349f90 service nova] Acquired lock "refresh_cache-7a72a1ab-3b3d-450d-a3a6-572b0035eb0e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.136510] env[63202]: DEBUG nova.network.neutron [req-a6f4a3f5-4a62-42c8-ba2d-34529a0cdc94 req-0260b0b6-7f95-45a5-878f-7daa4b349f90 service nova] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Refreshing network info cache for port aa2cbb5d-3724-4754-922b-da0798168b84 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 910.339306] env[63202]: INFO nova.compute.manager [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Took 22.75 seconds to build instance. [ 910.379445] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance b1bccea9-2d79-431a-8be0-0a5ab293542a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 910.432145] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquiring lock "7437595c-fa35-483e-95f3-b75405b6bd13" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.432514] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "7437595c-fa35-483e-95f3-b75405b6bd13" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.494799] env[63202]: DEBUG nova.network.neutron [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Successfully updated port: c7da6929-1bd1-4e2b-bcb8-609863cf6861 {{(pid=63202) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 910.506032] env[63202]: DEBUG oslo_vmware.api [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385453, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.546717] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385455, 'name': CreateVM_Task, 'duration_secs': 0.408047} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.546883] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 910.547615] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.547824] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.548211] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 910.548469] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f9fd9747-464f-4062-a8d4-b847050f8a5c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.553667] env[63202]: DEBUG oslo_vmware.api [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for the task: (returnval){ [ 910.553667] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52234c15-798a-6a7e-847b-da99ef02415d" [ 910.553667] env[63202]: _type = "Task" [ 910.553667] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.562755] env[63202]: DEBUG oslo_vmware.api [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52234c15-798a-6a7e-847b-da99ef02415d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.841803] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ec51fb29-0b87-4933-86c2-53b27f6f8251 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Lock "3d1518d3-bed2-4b7b-af1d-d7da49e92874" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 67.198s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.844279] env[63202]: DEBUG nova.network.neutron [req-a6f4a3f5-4a62-42c8-ba2d-34529a0cdc94 req-0260b0b6-7f95-45a5-878f-7daa4b349f90 service nova] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Updated VIF entry in instance network info cache for port aa2cbb5d-3724-4754-922b-da0798168b84. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 910.844585] env[63202]: DEBUG nova.network.neutron [req-a6f4a3f5-4a62-42c8-ba2d-34529a0cdc94 req-0260b0b6-7f95-45a5-878f-7daa4b349f90 service nova] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Updating instance_info_cache with network_info: [{"id": "aa2cbb5d-3724-4754-922b-da0798168b84", "address": "fa:16:3e:6b:99:b5", "network": {"id": "e6c760f6-ded9-4905-b303-dbcde0d037b9", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1433567753-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e914f87450949b1a39866e8cfa3a0eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "26472e27-9835-4f87-ab7f-ca24dfee4e83", "external-id": "nsx-vlan-transportzone-335", "segmentation_id": 335, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa2cbb5d-37", "ovs_interfaceid": "aa2cbb5d-3724-4754-922b-da0798168b84", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 910.882714] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance f3652744-e072-4700-80d4-b9eca414c5cb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 911.001337] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "refresh_cache-e775e5e1-521a-4fc7-80e6-bcb6a70516c5" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.001509] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquired lock "refresh_cache-e775e5e1-521a-4fc7-80e6-bcb6a70516c5" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.001677] env[63202]: DEBUG nova.network.neutron [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 911.008755] env[63202]: DEBUG oslo_vmware.api [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385453, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.547895} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.009018] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore2] 61a6528e-4e4e-49b7-be7b-dd30bd8023d8/61a6528e-4e4e-49b7-be7b-dd30bd8023d8.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 911.009225] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 911.009491] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-532a8ac5-d922-4ac4-bc85-2fd0c5bdc8ec {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.018481] env[63202]: DEBUG oslo_vmware.api [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Waiting for the task: (returnval){ [ 911.018481] env[63202]: value = "task-1385456" [ 911.018481] env[63202]: _type = "Task" [ 911.018481] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.027564] env[63202]: DEBUG oslo_vmware.api [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385456, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.067025] env[63202]: DEBUG oslo_vmware.api [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52234c15-798a-6a7e-847b-da99ef02415d, 'name': SearchDatastore_Task, 'duration_secs': 0.010559} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.067025] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.067025] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 911.067025] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.067277] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.067277] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 911.067277] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b4e867c1-c8ae-4fde-83ce-9308ede4ddc7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.076312] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 911.076731] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 911.077597] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-53bf3045-3d66-423b-9585-23598e944020 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.083180] env[63202]: DEBUG oslo_vmware.api [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for the task: (returnval){ [ 911.083180] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]526f9c0b-7d54-5325-0f97-2d1c3977c068" [ 911.083180] env[63202]: _type = "Task" [ 911.083180] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.093755] env[63202]: DEBUG oslo_vmware.api [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]526f9c0b-7d54-5325-0f97-2d1c3977c068, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.344937] env[63202]: DEBUG nova.compute.manager [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 911.347681] env[63202]: DEBUG oslo_concurrency.lockutils [req-a6f4a3f5-4a62-42c8-ba2d-34529a0cdc94 req-0260b0b6-7f95-45a5-878f-7daa4b349f90 service nova] Releasing lock "refresh_cache-7a72a1ab-3b3d-450d-a3a6-572b0035eb0e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.385857] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance fc7d3924-a624-4ea4-890c-7628595fb733 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 911.529449] env[63202]: DEBUG oslo_vmware.api [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385456, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.229093} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.529698] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 911.530561] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7262d6e1-51fe-45bf-bfaf-7baf007d72c6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.533544] env[63202]: DEBUG nova.network.neutron [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 911.555122] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] 61a6528e-4e4e-49b7-be7b-dd30bd8023d8/61a6528e-4e4e-49b7-be7b-dd30bd8023d8.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 911.557396] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-adfe0be0-871f-4471-bdab-be1c860a840d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.578600] env[63202]: DEBUG oslo_vmware.api [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Waiting for the task: (returnval){ [ 911.578600] env[63202]: value = "task-1385457" [ 911.578600] env[63202]: _type = "Task" [ 911.578600] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.587077] env[63202]: DEBUG oslo_vmware.api [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385457, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.595658] env[63202]: DEBUG oslo_vmware.api [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]526f9c0b-7d54-5325-0f97-2d1c3977c068, 'name': SearchDatastore_Task, 'duration_secs': 0.010292} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.596452] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c22cbb06-709f-46a4-a348-4de9c4a70412 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.605607] env[63202]: DEBUG oslo_vmware.api [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for the task: (returnval){ [ 911.605607] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]522b1693-0adf-386c-443a-7b2c6ac7a8c9" [ 911.605607] env[63202]: _type = "Task" [ 911.605607] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.614888] env[63202]: DEBUG oslo_vmware.api [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]522b1693-0adf-386c-443a-7b2c6ac7a8c9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.700752] env[63202]: DEBUG nova.network.neutron [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Updating instance_info_cache with network_info: [{"id": "c7da6929-1bd1-4e2b-bcb8-609863cf6861", "address": "fa:16:3e:cf:6c:83", "network": {"id": "18390479-16ce-4012-9ba6-abf19b5616d1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1376026898-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22352c25bca8416a948014391a5389ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "085fb0ff-9285-4f1d-a008-a14da4844357", "external-id": "nsx-vlan-transportzone-729", "segmentation_id": 729, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7da6929-1b", "ovs_interfaceid": "c7da6929-1bd1-4e2b-bcb8-609863cf6861", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.869760] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.888900] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance d0d6e380-9337-4f69-8434-6a33ac8a33d6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 912.089350] env[63202]: DEBUG oslo_vmware.api [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385457, 'name': ReconfigVM_Task, 'duration_secs': 0.289727} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.089629] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Reconfigured VM instance instance-00000046 to attach disk [datastore2] 61a6528e-4e4e-49b7-be7b-dd30bd8023d8/61a6528e-4e4e-49b7-be7b-dd30bd8023d8.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 912.090336] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-aa32212b-d230-4ce5-8885-24fc658e5248 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.098161] env[63202]: DEBUG oslo_vmware.api [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Waiting for the task: (returnval){ [ 912.098161] env[63202]: value = "task-1385458" [ 912.098161] env[63202]: _type = "Task" [ 912.098161] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.106663] env[63202]: DEBUG oslo_vmware.api [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385458, 'name': Rename_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.114816] env[63202]: DEBUG oslo_vmware.api [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]522b1693-0adf-386c-443a-7b2c6ac7a8c9, 'name': SearchDatastore_Task, 'duration_secs': 0.011513} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.115062] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.115321] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e/7a72a1ab-3b3d-450d-a3a6-572b0035eb0e.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 912.115572] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7a673229-6020-452a-ab7e-543f60098888 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.122065] env[63202]: DEBUG oslo_vmware.api [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for the task: (returnval){ [ 912.122065] env[63202]: value = "task-1385459" [ 912.122065] env[63202]: _type = "Task" [ 912.122065] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.130997] env[63202]: DEBUG oslo_vmware.api [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385459, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.173102] env[63202]: DEBUG nova.compute.manager [req-7583310b-52a8-42f0-9406-6669cc6f8c76 req-2b68f6c6-226a-4af6-969c-d63b1e51df8b service nova] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Received event network-vif-plugged-c7da6929-1bd1-4e2b-bcb8-609863cf6861 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 912.173385] env[63202]: DEBUG oslo_concurrency.lockutils [req-7583310b-52a8-42f0-9406-6669cc6f8c76 req-2b68f6c6-226a-4af6-969c-d63b1e51df8b service nova] Acquiring lock "e775e5e1-521a-4fc7-80e6-bcb6a70516c5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.173617] env[63202]: DEBUG oslo_concurrency.lockutils [req-7583310b-52a8-42f0-9406-6669cc6f8c76 req-2b68f6c6-226a-4af6-969c-d63b1e51df8b service nova] Lock "e775e5e1-521a-4fc7-80e6-bcb6a70516c5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.173777] env[63202]: DEBUG oslo_concurrency.lockutils [req-7583310b-52a8-42f0-9406-6669cc6f8c76 req-2b68f6c6-226a-4af6-969c-d63b1e51df8b service nova] Lock "e775e5e1-521a-4fc7-80e6-bcb6a70516c5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.173938] env[63202]: DEBUG nova.compute.manager [req-7583310b-52a8-42f0-9406-6669cc6f8c76 req-2b68f6c6-226a-4af6-969c-d63b1e51df8b service nova] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] No waiting events found dispatching network-vif-plugged-c7da6929-1bd1-4e2b-bcb8-609863cf6861 {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 912.174111] env[63202]: WARNING nova.compute.manager [req-7583310b-52a8-42f0-9406-6669cc6f8c76 req-2b68f6c6-226a-4af6-969c-d63b1e51df8b service nova] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Received unexpected event network-vif-plugged-c7da6929-1bd1-4e2b-bcb8-609863cf6861 for instance with vm_state building and task_state spawning. [ 912.174260] env[63202]: DEBUG nova.compute.manager [req-7583310b-52a8-42f0-9406-6669cc6f8c76 req-2b68f6c6-226a-4af6-969c-d63b1e51df8b service nova] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Received event network-changed-c7da6929-1bd1-4e2b-bcb8-609863cf6861 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 912.174407] env[63202]: DEBUG nova.compute.manager [req-7583310b-52a8-42f0-9406-6669cc6f8c76 req-2b68f6c6-226a-4af6-969c-d63b1e51df8b service nova] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Refreshing instance network info cache due to event network-changed-c7da6929-1bd1-4e2b-bcb8-609863cf6861. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 912.174567] env[63202]: DEBUG oslo_concurrency.lockutils [req-7583310b-52a8-42f0-9406-6669cc6f8c76 req-2b68f6c6-226a-4af6-969c-d63b1e51df8b service nova] Acquiring lock "refresh_cache-e775e5e1-521a-4fc7-80e6-bcb6a70516c5" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.203101] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Releasing lock "refresh_cache-e775e5e1-521a-4fc7-80e6-bcb6a70516c5" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.203454] env[63202]: DEBUG nova.compute.manager [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Instance network_info: |[{"id": "c7da6929-1bd1-4e2b-bcb8-609863cf6861", "address": "fa:16:3e:cf:6c:83", "network": {"id": "18390479-16ce-4012-9ba6-abf19b5616d1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1376026898-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22352c25bca8416a948014391a5389ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "085fb0ff-9285-4f1d-a008-a14da4844357", "external-id": "nsx-vlan-transportzone-729", "segmentation_id": 729, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7da6929-1b", "ovs_interfaceid": "c7da6929-1bd1-4e2b-bcb8-609863cf6861", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 912.203716] env[63202]: DEBUG oslo_concurrency.lockutils [req-7583310b-52a8-42f0-9406-6669cc6f8c76 req-2b68f6c6-226a-4af6-969c-d63b1e51df8b service nova] Acquired lock "refresh_cache-e775e5e1-521a-4fc7-80e6-bcb6a70516c5" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.203901] env[63202]: DEBUG nova.network.neutron [req-7583310b-52a8-42f0-9406-6669cc6f8c76 req-2b68f6c6-226a-4af6-969c-d63b1e51df8b service nova] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Refreshing network info cache for port c7da6929-1bd1-4e2b-bcb8-609863cf6861 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 912.205208] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cf:6c:83', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '085fb0ff-9285-4f1d-a008-a14da4844357', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c7da6929-1bd1-4e2b-bcb8-609863cf6861', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 912.213173] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Creating folder: Project (22352c25bca8416a948014391a5389ac). Parent ref: group-v294090. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 912.216496] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bcb6cb62-9cca-40f7-8ac6-dce6537e6110 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.231280] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Created folder: Project (22352c25bca8416a948014391a5389ac) in parent group-v294090. [ 912.231487] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Creating folder: Instances. Parent ref: group-v294138. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 912.231779] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-268e36f9-3b69-489c-a752-e8ef847e272c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.243873] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Created folder: Instances in parent group-v294138. [ 912.244193] env[63202]: DEBUG oslo.service.loopingcall [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 912.244406] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 912.244615] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f2d11e71-74e9-42c4-ad0f-4631eca04480 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.267130] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 912.267130] env[63202]: value = "task-1385462" [ 912.267130] env[63202]: _type = "Task" [ 912.267130] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.280024] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385462, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.392838] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 8755bceb-d510-4429-bd98-d6a63faf739d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 912.454168] env[63202]: DEBUG nova.network.neutron [req-7583310b-52a8-42f0-9406-6669cc6f8c76 req-2b68f6c6-226a-4af6-969c-d63b1e51df8b service nova] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Updated VIF entry in instance network info cache for port c7da6929-1bd1-4e2b-bcb8-609863cf6861. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 912.454603] env[63202]: DEBUG nova.network.neutron [req-7583310b-52a8-42f0-9406-6669cc6f8c76 req-2b68f6c6-226a-4af6-969c-d63b1e51df8b service nova] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Updating instance_info_cache with network_info: [{"id": "c7da6929-1bd1-4e2b-bcb8-609863cf6861", "address": "fa:16:3e:cf:6c:83", "network": {"id": "18390479-16ce-4012-9ba6-abf19b5616d1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1376026898-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22352c25bca8416a948014391a5389ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "085fb0ff-9285-4f1d-a008-a14da4844357", "external-id": "nsx-vlan-transportzone-729", "segmentation_id": 729, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7da6929-1b", "ovs_interfaceid": "c7da6929-1bd1-4e2b-bcb8-609863cf6861", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.609509] env[63202]: DEBUG oslo_vmware.api [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385458, 'name': Rename_Task, 'duration_secs': 0.154405} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.609828] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 912.610159] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-03bf8bdd-1f23-43bd-acb2-940abe28f97b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.618272] env[63202]: DEBUG oslo_vmware.api [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Waiting for the task: (returnval){ [ 912.618272] env[63202]: value = "task-1385463" [ 912.618272] env[63202]: _type = "Task" [ 912.618272] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.629841] env[63202]: DEBUG oslo_vmware.api [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385463, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.635862] env[63202]: DEBUG oslo_vmware.api [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385459, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.500286} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.636139] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e/7a72a1ab-3b3d-450d-a3a6-572b0035eb0e.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 912.636350] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 912.636611] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-08c7b6b8-01da-4b01-8dce-b245c315b6a3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.645529] env[63202]: DEBUG oslo_vmware.api [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for the task: (returnval){ [ 912.645529] env[63202]: value = "task-1385464" [ 912.645529] env[63202]: _type = "Task" [ 912.645529] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.656973] env[63202]: DEBUG oslo_vmware.api [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385464, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.778137] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385462, 'name': CreateVM_Task, 'duration_secs': 0.496909} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.778306] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 912.779020] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.779212] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.779558] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 912.779831] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ba877ad3-55cd-41fe-b6b0-a38e74ac06d8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.784922] env[63202]: DEBUG oslo_vmware.api [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for the task: (returnval){ [ 912.784922] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52f4ca71-403b-c47c-2a88-36f2ae9ae80c" [ 912.784922] env[63202]: _type = "Task" [ 912.784922] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.792888] env[63202]: DEBUG oslo_vmware.api [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52f4ca71-403b-c47c-2a88-36f2ae9ae80c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.896023] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance f3128c09-3680-4b0b-b463-3d6cd203fcf4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 912.896405] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Total usable vcpus: 48, total allocated vcpus: 7 {{(pid=63202) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 912.896563] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1920MB phys_disk=200GB used_disk=7GB total_vcpus=48 used_vcpus=7 pci_stats=[] {{(pid=63202) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 912.961065] env[63202]: DEBUG oslo_concurrency.lockutils [req-7583310b-52a8-42f0-9406-6669cc6f8c76 req-2b68f6c6-226a-4af6-969c-d63b1e51df8b service nova] Releasing lock "refresh_cache-e775e5e1-521a-4fc7-80e6-bcb6a70516c5" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.130403] env[63202]: DEBUG oslo_vmware.api [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385463, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.146981] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a71cfeeb-836c-4e4a-a17f-0637cf2b4fae {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.160711] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccc9253e-d09c-4a8c-997c-0943c65f97e2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.163731] env[63202]: DEBUG oslo_vmware.api [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385464, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079499} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.163980] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 913.165020] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5320319d-c010-47f8-8bb7-49b20e9635ec {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.192860] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5fcee76-1c41-416b-b76c-773a71b9313c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.212683] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Reconfiguring VM instance instance-00000047 to attach disk [datastore1] 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e/7a72a1ab-3b3d-450d-a3a6-572b0035eb0e.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 913.213352] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c6b615cf-ee36-4fb7-b593-6bbd00f22f00 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.231953] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bc02fb2-ce55-4f92-a02a-985232e208e5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.236119] env[63202]: DEBUG oslo_vmware.api [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for the task: (returnval){ [ 913.236119] env[63202]: value = "task-1385465" [ 913.236119] env[63202]: _type = "Task" [ 913.236119] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.247088] env[63202]: DEBUG nova.compute.provider_tree [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 913.254144] env[63202]: DEBUG oslo_vmware.api [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385465, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.296386] env[63202]: DEBUG oslo_vmware.api [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52f4ca71-403b-c47c-2a88-36f2ae9ae80c, 'name': SearchDatastore_Task, 'duration_secs': 0.022621} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.296667] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.296982] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 913.297225] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.297323] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.297493] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 913.297770] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-abbedf96-7c1a-4e82-bf78-4fa4a68879f7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.309112] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 913.309249] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 913.310039] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bfb645e9-b9c9-4b62-bd92-c33a1f3314aa {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.316306] env[63202]: DEBUG oslo_vmware.api [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for the task: (returnval){ [ 913.316306] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]529f698b-11c7-901e-d638-0712e6776df9" [ 913.316306] env[63202]: _type = "Task" [ 913.316306] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.325526] env[63202]: DEBUG oslo_vmware.api [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]529f698b-11c7-901e-d638-0712e6776df9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.632082] env[63202]: DEBUG oslo_vmware.api [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385463, 'name': PowerOnVM_Task, 'duration_secs': 0.516649} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.632082] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 913.632082] env[63202]: INFO nova.compute.manager [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Took 8.70 seconds to spawn the instance on the hypervisor. [ 913.632478] env[63202]: DEBUG nova.compute.manager [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 913.633150] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-099c28a7-fe70-4b34-96d7-ea8d063bd219 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.747352] env[63202]: DEBUG oslo_vmware.api [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385465, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.751638] env[63202]: DEBUG nova.scheduler.client.report [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 913.826084] env[63202]: DEBUG oslo_vmware.api [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]529f698b-11c7-901e-d638-0712e6776df9, 'name': SearchDatastore_Task, 'duration_secs': 0.022027} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.826891] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f95141d-bc9e-407d-b97f-110a717a0f42 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.832190] env[63202]: DEBUG oslo_vmware.api [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for the task: (returnval){ [ 913.832190] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]5293b27d-0ded-b190-03df-2fc133399d4b" [ 913.832190] env[63202]: _type = "Task" [ 913.832190] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.840084] env[63202]: DEBUG oslo_vmware.api [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]5293b27d-0ded-b190-03df-2fc133399d4b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.154761] env[63202]: INFO nova.compute.manager [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Took 23.44 seconds to build instance. [ 914.248209] env[63202]: DEBUG oslo_vmware.api [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385465, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.255511] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63202) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 914.255818] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 6.462s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.255965] env[63202]: DEBUG oslo_concurrency.lockutils [None req-0915f74e-cfc6-45f1-9f44-9af3c69affef tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.309s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.256157] env[63202]: DEBUG oslo_concurrency.lockutils [None req-0915f74e-cfc6-45f1-9f44-9af3c69affef tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.258163] env[63202]: DEBUG oslo_concurrency.lockutils [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.599s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.259721] env[63202]: INFO nova.compute.claims [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 914.285467] env[63202]: INFO nova.scheduler.client.report [None req-0915f74e-cfc6-45f1-9f44-9af3c69affef tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Deleted allocations for instance a800fa5d-46a1-4a96-8ea2-26e34b869aa6 [ 914.343679] env[63202]: DEBUG oslo_vmware.api [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]5293b27d-0ded-b190-03df-2fc133399d4b, 'name': SearchDatastore_Task, 'duration_secs': 0.009396} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.343804] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.344223] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] e775e5e1-521a-4fc7-80e6-bcb6a70516c5/e775e5e1-521a-4fc7-80e6-bcb6a70516c5.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 914.344522] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3f5ff0e2-ea9e-4e9b-b099-c4681fa8773b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.352741] env[63202]: DEBUG oslo_vmware.api [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for the task: (returnval){ [ 914.352741] env[63202]: value = "task-1385466" [ 914.352741] env[63202]: _type = "Task" [ 914.352741] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.361637] env[63202]: DEBUG oslo_vmware.api [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385466, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.656856] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ac2bd3e6-cabd-4e2c-8f72-6459d6cbecc0 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Lock "61a6528e-4e4e-49b7-be7b-dd30bd8023d8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.697s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.751448] env[63202]: DEBUG oslo_vmware.api [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385465, 'name': ReconfigVM_Task, 'duration_secs': 1.170026} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.751739] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Reconfigured VM instance instance-00000047 to attach disk [datastore1] 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e/7a72a1ab-3b3d-450d-a3a6-572b0035eb0e.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 914.752434] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e2150dd3-47aa-40a4-a7ea-c35776431350 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.762203] env[63202]: DEBUG oslo_vmware.api [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for the task: (returnval){ [ 914.762203] env[63202]: value = "task-1385467" [ 914.762203] env[63202]: _type = "Task" [ 914.762203] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.776924] env[63202]: DEBUG oslo_vmware.api [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385467, 'name': Rename_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.794484] env[63202]: DEBUG oslo_concurrency.lockutils [None req-0915f74e-cfc6-45f1-9f44-9af3c69affef tempest-ServersAaction247Test-2045226228 tempest-ServersAaction247Test-2045226228-project-member] Lock "a800fa5d-46a1-4a96-8ea2-26e34b869aa6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.619s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.863958] env[63202]: DEBUG oslo_vmware.api [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385466, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.159649] env[63202]: DEBUG nova.compute.manager [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 915.262495] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 915.276555] env[63202]: DEBUG oslo_vmware.api [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385467, 'name': Rename_Task, 'duration_secs': 0.189435} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.277594] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 915.277594] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8083b52a-4a37-4fbc-bea1-936e6a31203e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.285237] env[63202]: DEBUG oslo_vmware.api [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for the task: (returnval){ [ 915.285237] env[63202]: value = "task-1385468" [ 915.285237] env[63202]: _type = "Task" [ 915.285237] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.298575] env[63202]: DEBUG oslo_vmware.api [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385468, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.370766] env[63202]: DEBUG oslo_vmware.api [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385466, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.519565} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.371147] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] e775e5e1-521a-4fc7-80e6-bcb6a70516c5/e775e5e1-521a-4fc7-80e6-bcb6a70516c5.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 915.371517] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 915.372081] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7fc655b2-98f8-4110-95f6-f1bfc4522083 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.383830] env[63202]: DEBUG oslo_vmware.api [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for the task: (returnval){ [ 915.383830] env[63202]: value = "task-1385469" [ 915.383830] env[63202]: _type = "Task" [ 915.383830] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.396693] env[63202]: DEBUG oslo_vmware.api [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385469, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.548704] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebf295f2-822a-4599-bb82-d25ce2379e14 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.557109] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9635d166-604d-4518-ad94-36b68447b990 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.590492] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47bf24cc-f0c2-4d7a-8213-f7434dbf6bf5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.599266] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53912db6-5061-42fd-a232-190daa2da12e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.615723] env[63202]: DEBUG nova.compute.provider_tree [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 915.683625] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.772846] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 915.772846] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Starting heal instance info cache {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 915.772949] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Rebuilding the list of instances to heal {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 915.797328] env[63202]: DEBUG oslo_vmware.api [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385468, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.894742] env[63202]: DEBUG oslo_vmware.api [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385469, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.242883} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.895128] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 915.895955] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4f4784b-b008-4ce9-8a72-d2f3c2489446 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.924879] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Reconfiguring VM instance instance-00000048 to attach disk [datastore1] e775e5e1-521a-4fc7-80e6-bcb6a70516c5/e775e5e1-521a-4fc7-80e6-bcb6a70516c5.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 915.925231] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-45dae425-7016-4052-a747-5e5b03403aaa {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.947607] env[63202]: DEBUG oslo_vmware.api [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for the task: (returnval){ [ 915.947607] env[63202]: value = "task-1385470" [ 915.947607] env[63202]: _type = "Task" [ 915.947607] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.956955] env[63202]: DEBUG oslo_vmware.api [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385470, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.119644] env[63202]: DEBUG nova.scheduler.client.report [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 916.277177] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Skipping network cache update for instance because it is Building. {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 916.277706] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Skipping network cache update for instance because it is Building. {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 916.277706] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Skipping network cache update for instance because it is Building. {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 916.297104] env[63202]: DEBUG oslo_vmware.api [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385468, 'name': PowerOnVM_Task, 'duration_secs': 0.710455} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.297104] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 916.297104] env[63202]: INFO nova.compute.manager [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Took 8.89 seconds to spawn the instance on the hypervisor. [ 916.297104] env[63202]: DEBUG nova.compute.manager [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 916.297835] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9350c3f-14f9-448e-9f47-0bdf9c1bd8f6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.348078] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Acquiring lock "refresh_cache-0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.348271] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Acquired lock "refresh_cache-0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.348662] env[63202]: DEBUG nova.network.neutron [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Forcefully refreshing network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 916.348662] env[63202]: DEBUG nova.objects.instance [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lazy-loading 'info_cache' on Instance uuid 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 916.459416] env[63202]: DEBUG oslo_vmware.api [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385470, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.626358] env[63202]: DEBUG oslo_concurrency.lockutils [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.368s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.626978] env[63202]: DEBUG nova.compute.manager [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 916.629837] env[63202]: DEBUG oslo_concurrency.lockutils [None req-50e06c6a-8e1f-4273-ad90-21c08bf53c81 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.630144] env[63202]: DEBUG oslo_concurrency.lockutils [None req-50e06c6a-8e1f-4273-ad90-21c08bf53c81 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.632237] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.071s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.633748] env[63202]: INFO nova.compute.claims [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 916.661390] env[63202]: INFO nova.scheduler.client.report [None req-50e06c6a-8e1f-4273-ad90-21c08bf53c81 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Deleted allocations for instance 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a [ 916.772276] env[63202]: DEBUG oslo_concurrency.lockutils [None req-26047fad-f98a-4024-beda-18082bb678ed tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Acquiring lock "36291628-af7d-43cf-8149-09f57df47890" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.772642] env[63202]: DEBUG oslo_concurrency.lockutils [None req-26047fad-f98a-4024-beda-18082bb678ed tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Lock "36291628-af7d-43cf-8149-09f57df47890" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.772790] env[63202]: DEBUG nova.compute.manager [None req-26047fad-f98a-4024-beda-18082bb678ed tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 916.773962] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd2c10ca-2553-4366-b88c-25af8d301bf5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.781869] env[63202]: DEBUG nova.compute.manager [None req-26047fad-f98a-4024-beda-18082bb678ed tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63202) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 916.782427] env[63202]: DEBUG nova.objects.instance [None req-26047fad-f98a-4024-beda-18082bb678ed tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Lazy-loading 'flavor' on Instance uuid 36291628-af7d-43cf-8149-09f57df47890 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 916.819238] env[63202]: INFO nova.compute.manager [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Took 24.52 seconds to build instance. [ 916.963749] env[63202]: DEBUG oslo_vmware.api [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385470, 'name': ReconfigVM_Task, 'duration_secs': 0.736202} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.964195] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Reconfigured VM instance instance-00000048 to attach disk [datastore1] e775e5e1-521a-4fc7-80e6-bcb6a70516c5/e775e5e1-521a-4fc7-80e6-bcb6a70516c5.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 916.965037] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0789fb54-7ed5-4b2e-aa92-b2f49362f359 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.973812] env[63202]: DEBUG oslo_vmware.api [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for the task: (returnval){ [ 916.973812] env[63202]: value = "task-1385471" [ 916.973812] env[63202]: _type = "Task" [ 916.973812] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.988034] env[63202]: DEBUG oslo_vmware.api [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385471, 'name': Rename_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.139366] env[63202]: DEBUG nova.compute.utils [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 917.144057] env[63202]: DEBUG nova.compute.manager [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 917.144057] env[63202]: DEBUG nova.network.neutron [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 917.170373] env[63202]: DEBUG oslo_concurrency.lockutils [None req-50e06c6a-8e1f-4273-ad90-21c08bf53c81 tempest-DeleteServersAdminTestJSON-1736901598 tempest-DeleteServersAdminTestJSON-1736901598-project-member] Lock "5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.484s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.190109] env[63202]: DEBUG nova.policy [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8d1ac6136a7f42f0ada077ec276320ac', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '90f7aa97c6a84ee2bd79518f6e7e2ad3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 917.288822] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-26047fad-f98a-4024-beda-18082bb678ed tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 917.289352] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b2e259cd-f1d0-455b-8704-7bb06928b111 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.296965] env[63202]: DEBUG oslo_vmware.api [None req-26047fad-f98a-4024-beda-18082bb678ed tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Waiting for the task: (returnval){ [ 917.296965] env[63202]: value = "task-1385472" [ 917.296965] env[63202]: _type = "Task" [ 917.296965] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.306536] env[63202]: DEBUG oslo_vmware.api [None req-26047fad-f98a-4024-beda-18082bb678ed tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385472, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.321375] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d853c7f2-55eb-46d1-a69e-667b5065b500 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lock "7a72a1ab-3b3d-450d-a3a6-572b0035eb0e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.665s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.487160] env[63202]: DEBUG oslo_vmware.api [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385471, 'name': Rename_Task, 'duration_secs': 0.279953} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.487160] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 917.487526] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f72ad9a3-3d5c-4ba2-bc21-141ce05973f7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.496149] env[63202]: DEBUG oslo_vmware.api [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for the task: (returnval){ [ 917.496149] env[63202]: value = "task-1385473" [ 917.496149] env[63202]: _type = "Task" [ 917.496149] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.506680] env[63202]: DEBUG oslo_vmware.api [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385473, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.607034] env[63202]: DEBUG nova.network.neutron [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Successfully created port: 9a8ab9d1-003d-492b-9a0b-726c69f6df9d {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 917.644964] env[63202]: DEBUG nova.compute.manager [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 917.807260] env[63202]: DEBUG oslo_vmware.api [None req-26047fad-f98a-4024-beda-18082bb678ed tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385472, 'name': PowerOffVM_Task, 'duration_secs': 0.360449} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.807694] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-26047fad-f98a-4024-beda-18082bb678ed tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 917.807989] env[63202]: DEBUG nova.compute.manager [None req-26047fad-f98a-4024-beda-18082bb678ed tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 917.809332] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c1b100c-eabc-4eec-a23f-36aaaa2abe51 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.825891] env[63202]: DEBUG nova.compute.manager [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 917.953228] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ba37bb0-943f-4e12-a616-bd6b13966bd8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.962110] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a70e1d3b-6150-4537-afb4-4a35fe23e5b1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.997722] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ef2f3f9-6f51-4fa8-a0e4-73f5c8befeb4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.011947] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd35a979-4e23-41ec-b2e0-58ec539e7234 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.015635] env[63202]: DEBUG oslo_vmware.api [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385473, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.027022] env[63202]: DEBUG nova.compute.provider_tree [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 918.157757] env[63202]: DEBUG nova.network.neutron [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Updating instance_info_cache with network_info: [{"id": "e5621970-d0a8-4803-94ef-f6d577bdf6d2", "address": "fa:16:3e:00:ba:72", "network": {"id": "345538f4-e791-4fc6-9719-237c11f1382a", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-449651008-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.170", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a9e939cdf4a94a8aa0107a5761771c2c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa09e855-8af1-419b-b78d-8ffcc94b1bfb", "external-id": "nsx-vlan-transportzone-901", "segmentation_id": 901, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape5621970-d0", "ovs_interfaceid": "e5621970-d0a8-4803-94ef-f6d577bdf6d2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.246932] env[63202]: DEBUG nova.compute.manager [None req-5423f696-c38b-4f13-988f-bd3e2591037c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 918.247998] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ab571e4-49b0-464e-85ed-c123addaad7e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.324711] env[63202]: DEBUG oslo_concurrency.lockutils [None req-26047fad-f98a-4024-beda-18082bb678ed tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Lock "36291628-af7d-43cf-8149-09f57df47890" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.552s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.347534] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.515424] env[63202]: DEBUG oslo_vmware.api [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385473, 'name': PowerOnVM_Task, 'duration_secs': 0.922677} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.515760] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 918.516033] env[63202]: INFO nova.compute.manager [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Took 9.17 seconds to spawn the instance on the hypervisor. [ 918.516578] env[63202]: DEBUG nova.compute.manager [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 918.517101] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba88592a-0959-4a12-8f2a-5999a20b633c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.529578] env[63202]: DEBUG nova.scheduler.client.report [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 918.660741] env[63202]: DEBUG nova.compute.manager [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 918.664287] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Releasing lock "refresh_cache-0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.664549] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Updated the network info_cache for instance {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 918.664802] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 918.688792] env[63202]: DEBUG nova.virt.hardware [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 918.689162] env[63202]: DEBUG nova.virt.hardware [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 918.689377] env[63202]: DEBUG nova.virt.hardware [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 918.689615] env[63202]: DEBUG nova.virt.hardware [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 918.689806] env[63202]: DEBUG nova.virt.hardware [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 918.689998] env[63202]: DEBUG nova.virt.hardware [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 918.690304] env[63202]: DEBUG nova.virt.hardware [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 918.690536] env[63202]: DEBUG nova.virt.hardware [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 918.690784] env[63202]: DEBUG nova.virt.hardware [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 918.691016] env[63202]: DEBUG nova.virt.hardware [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 918.691290] env[63202]: DEBUG nova.virt.hardware [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 918.692555] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d8fab38-6682-4a3c-903e-207453d57aba {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.704649] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcd34538-f637-4d51-9d4c-ff0910c495a5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.759886] env[63202]: INFO nova.compute.manager [None req-5423f696-c38b-4f13-988f-bd3e2591037c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] instance snapshotting [ 918.765905] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9adfc8b9-02d0-412f-b633-0690ccf3a4f9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.787390] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fc07f07-3bf5-4089-affe-1addf6d3f536 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.036234] env[63202]: INFO nova.compute.manager [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Took 26.31 seconds to build instance. [ 919.036234] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.404s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.036234] env[63202]: DEBUG nova.compute.manager [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 919.038889] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.393s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.040242] env[63202]: INFO nova.compute.claims [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 919.302020] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-5423f696-c38b-4f13-988f-bd3e2591037c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Creating Snapshot of the VM instance {{(pid=63202) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 919.302020] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-5d3b3ba9-56fe-40fb-a105-1f7f477a8bdc {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.308360] env[63202]: DEBUG oslo_vmware.api [None req-5423f696-c38b-4f13-988f-bd3e2591037c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for the task: (returnval){ [ 919.308360] env[63202]: value = "task-1385474" [ 919.308360] env[63202]: _type = "Task" [ 919.308360] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.318358] env[63202]: DEBUG oslo_vmware.api [None req-5423f696-c38b-4f13-988f-bd3e2591037c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385474, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.320882] env[63202]: DEBUG nova.compute.manager [req-524b59cd-1204-4684-9740-48b841095d13 req-24974185-c828-4ed9-a52e-9af7b1251062 service nova] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Received event network-vif-plugged-9a8ab9d1-003d-492b-9a0b-726c69f6df9d {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 919.321151] env[63202]: DEBUG oslo_concurrency.lockutils [req-524b59cd-1204-4684-9740-48b841095d13 req-24974185-c828-4ed9-a52e-9af7b1251062 service nova] Acquiring lock "753d190b-f4a4-4438-bc98-94564e3ec73d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.321390] env[63202]: DEBUG oslo_concurrency.lockutils [req-524b59cd-1204-4684-9740-48b841095d13 req-24974185-c828-4ed9-a52e-9af7b1251062 service nova] Lock "753d190b-f4a4-4438-bc98-94564e3ec73d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.321539] env[63202]: DEBUG oslo_concurrency.lockutils [req-524b59cd-1204-4684-9740-48b841095d13 req-24974185-c828-4ed9-a52e-9af7b1251062 service nova] Lock "753d190b-f4a4-4438-bc98-94564e3ec73d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.321727] env[63202]: DEBUG nova.compute.manager [req-524b59cd-1204-4684-9740-48b841095d13 req-24974185-c828-4ed9-a52e-9af7b1251062 service nova] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] No waiting events found dispatching network-vif-plugged-9a8ab9d1-003d-492b-9a0b-726c69f6df9d {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 919.321900] env[63202]: WARNING nova.compute.manager [req-524b59cd-1204-4684-9740-48b841095d13 req-24974185-c828-4ed9-a52e-9af7b1251062 service nova] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Received unexpected event network-vif-plugged-9a8ab9d1-003d-492b-9a0b-726c69f6df9d for instance with vm_state building and task_state spawning. [ 919.325120] env[63202]: DEBUG nova.objects.instance [None req-12584d87-b37a-41e6-b8aa-fb8129924ec5 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Lazy-loading 'flavor' on Instance uuid 36291628-af7d-43cf-8149-09f57df47890 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 919.540668] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2f4c677d-7999-4e4d-ac36-3d699c987428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "e775e5e1-521a-4fc7-80e6-bcb6a70516c5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.117s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.545111] env[63202]: DEBUG nova.compute.utils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 919.552262] env[63202]: DEBUG nova.compute.manager [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 919.552262] env[63202]: DEBUG nova.network.neutron [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 919.626269] env[63202]: DEBUG nova.policy [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9c4413afc60d4bf6aa0dd3a08b588b66', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'af1232a17bfd4bd3a3fe860f7da83994', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 919.772200] env[63202]: DEBUG nova.network.neutron [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Successfully updated port: 9a8ab9d1-003d-492b-9a0b-726c69f6df9d {{(pid=63202) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 919.805829] env[63202]: DEBUG nova.compute.manager [req-717e4253-2815-4950-ab5e-402b0b187dc9 req-ef2a5342-3673-46fe-aa57-21289498ff2e service nova] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Received event network-changed-9a8ab9d1-003d-492b-9a0b-726c69f6df9d {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 919.806036] env[63202]: DEBUG nova.compute.manager [req-717e4253-2815-4950-ab5e-402b0b187dc9 req-ef2a5342-3673-46fe-aa57-21289498ff2e service nova] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Refreshing instance network info cache due to event network-changed-9a8ab9d1-003d-492b-9a0b-726c69f6df9d. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 919.806248] env[63202]: DEBUG oslo_concurrency.lockutils [req-717e4253-2815-4950-ab5e-402b0b187dc9 req-ef2a5342-3673-46fe-aa57-21289498ff2e service nova] Acquiring lock "refresh_cache-753d190b-f4a4-4438-bc98-94564e3ec73d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.807248] env[63202]: DEBUG oslo_concurrency.lockutils [req-717e4253-2815-4950-ab5e-402b0b187dc9 req-ef2a5342-3673-46fe-aa57-21289498ff2e service nova] Acquired lock "refresh_cache-753d190b-f4a4-4438-bc98-94564e3ec73d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.807470] env[63202]: DEBUG nova.network.neutron [req-717e4253-2815-4950-ab5e-402b0b187dc9 req-ef2a5342-3673-46fe-aa57-21289498ff2e service nova] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Refreshing network info cache for port 9a8ab9d1-003d-492b-9a0b-726c69f6df9d {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 919.820455] env[63202]: DEBUG oslo_vmware.api [None req-5423f696-c38b-4f13-988f-bd3e2591037c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385474, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.829819] env[63202]: DEBUG oslo_concurrency.lockutils [None req-12584d87-b37a-41e6-b8aa-fb8129924ec5 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Acquiring lock "refresh_cache-36291628-af7d-43cf-8149-09f57df47890" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.829996] env[63202]: DEBUG oslo_concurrency.lockutils [None req-12584d87-b37a-41e6-b8aa-fb8129924ec5 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Acquired lock "refresh_cache-36291628-af7d-43cf-8149-09f57df47890" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.830210] env[63202]: DEBUG nova.network.neutron [None req-12584d87-b37a-41e6-b8aa-fb8129924ec5 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 919.830523] env[63202]: DEBUG nova.objects.instance [None req-12584d87-b37a-41e6-b8aa-fb8129924ec5 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Lazy-loading 'info_cache' on Instance uuid 36291628-af7d-43cf-8149-09f57df47890 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 920.050499] env[63202]: DEBUG nova.compute.manager [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 920.079591] env[63202]: DEBUG nova.network.neutron [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Successfully created port: f4c43c2b-ec12-47f7-84d8-4190afbf1b7d {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 920.276602] env[63202]: DEBUG oslo_concurrency.lockutils [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Acquiring lock "refresh_cache-753d190b-f4a4-4438-bc98-94564e3ec73d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.321923] env[63202]: DEBUG oslo_vmware.api [None req-5423f696-c38b-4f13-988f-bd3e2591037c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385474, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.323602] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2f4f4cf-c3de-4bf6-8250-e8a9de53b5ee {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.330550] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57d1a5d9-ae22-4ac6-9d81-50c074d50440 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.335171] env[63202]: DEBUG nova.objects.base [None req-12584d87-b37a-41e6-b8aa-fb8129924ec5 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Object Instance<36291628-af7d-43cf-8149-09f57df47890> lazy-loaded attributes: flavor,info_cache {{(pid=63202) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 920.370130] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57f8b2cb-146b-4efd-9643-f0564d8b94e1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.374489] env[63202]: DEBUG nova.network.neutron [req-717e4253-2815-4950-ab5e-402b0b187dc9 req-ef2a5342-3673-46fe-aa57-21289498ff2e service nova] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 920.380275] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da7b8753-ac30-4736-baac-cb0f6e03fd2e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.398745] env[63202]: DEBUG nova.compute.provider_tree [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 920.554400] env[63202]: DEBUG nova.network.neutron [req-717e4253-2815-4950-ab5e-402b0b187dc9 req-ef2a5342-3673-46fe-aa57-21289498ff2e service nova] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.820028] env[63202]: DEBUG oslo_vmware.api [None req-5423f696-c38b-4f13-988f-bd3e2591037c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385474, 'name': CreateSnapshot_Task, 'duration_secs': 1.330635} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.820803] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-5423f696-c38b-4f13-988f-bd3e2591037c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Created Snapshot of the VM instance {{(pid=63202) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 920.822099] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-849b1618-8902-47ec-a7c5-b866e1d04f24 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.901582] env[63202]: DEBUG nova.scheduler.client.report [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 920.950030] env[63202]: DEBUG nova.compute.manager [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Stashing vm_state: active {{(pid=63202) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 921.061503] env[63202]: DEBUG oslo_concurrency.lockutils [req-717e4253-2815-4950-ab5e-402b0b187dc9 req-ef2a5342-3673-46fe-aa57-21289498ff2e service nova] Releasing lock "refresh_cache-753d190b-f4a4-4438-bc98-94564e3ec73d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.065023] env[63202]: DEBUG nova.compute.manager [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 921.065253] env[63202]: DEBUG oslo_concurrency.lockutils [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Acquired lock "refresh_cache-753d190b-f4a4-4438-bc98-94564e3ec73d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.065534] env[63202]: DEBUG nova.network.neutron [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 921.075091] env[63202]: DEBUG nova.network.neutron [None req-12584d87-b37a-41e6-b8aa-fb8129924ec5 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] Updating instance_info_cache with network_info: [{"id": "355608b4-3a2c-4f49-8b9a-6b379a97934d", "address": "fa:16:3e:59:3b:cb", "network": {"id": "d83c3dc7-a0c9-4401-8b08-99d493f23a8b", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-482332878-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "98fbc2f9559148b8b0e2a2136152f13c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap355608b4-3a", "ovs_interfaceid": "355608b4-3a2c-4f49-8b9a-6b379a97934d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.087379] env[63202]: DEBUG nova.virt.hardware [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 921.087628] env[63202]: DEBUG nova.virt.hardware [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 921.087780] env[63202]: DEBUG nova.virt.hardware [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 921.087956] env[63202]: DEBUG nova.virt.hardware [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 921.088113] env[63202]: DEBUG nova.virt.hardware [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 921.088264] env[63202]: DEBUG nova.virt.hardware [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 921.088461] env[63202]: DEBUG nova.virt.hardware [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 921.088612] env[63202]: DEBUG nova.virt.hardware [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 921.088773] env[63202]: DEBUG nova.virt.hardware [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 921.088930] env[63202]: DEBUG nova.virt.hardware [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 921.089588] env[63202]: DEBUG nova.virt.hardware [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 921.090741] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dd179b4-a3e3-4f44-9143-3e0aef75de96 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.100176] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fedae439-f4e2-4033-8670-4ae6b9dbffea {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.343616] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-5423f696-c38b-4f13-988f-bd3e2591037c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Creating linked-clone VM from snapshot {{(pid=63202) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 921.343945] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-d0910b41-8be7-47a3-9aca-3bad4898e826 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.353246] env[63202]: DEBUG oslo_vmware.api [None req-5423f696-c38b-4f13-988f-bd3e2591037c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for the task: (returnval){ [ 921.353246] env[63202]: value = "task-1385475" [ 921.353246] env[63202]: _type = "Task" [ 921.353246] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.362339] env[63202]: DEBUG oslo_vmware.api [None req-5423f696-c38b-4f13-988f-bd3e2591037c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385475, 'name': CloneVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.406415] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.367s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.408059] env[63202]: DEBUG nova.compute.manager [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 921.410967] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.921s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.412382] env[63202]: INFO nova.compute.claims [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 921.469316] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.577688] env[63202]: DEBUG oslo_concurrency.lockutils [None req-12584d87-b37a-41e6-b8aa-fb8129924ec5 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Releasing lock "refresh_cache-36291628-af7d-43cf-8149-09f57df47890" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.603998] env[63202]: DEBUG nova.network.neutron [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 921.797449] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Acquiring lock "a961b2a0-39c1-4267-9229-068e2b6ecd67" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.797707] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Lock "a961b2a0-39c1-4267-9229-068e2b6ecd67" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.827096] env[63202]: DEBUG nova.network.neutron [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Updating instance_info_cache with network_info: [{"id": "9a8ab9d1-003d-492b-9a0b-726c69f6df9d", "address": "fa:16:3e:62:14:16", "network": {"id": "ee6a8ee1-64c0-49e0-a3cb-d4df94f7c6e8", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1816904884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "90f7aa97c6a84ee2bd79518f6e7e2ad3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b107fab-ee71-47db-ad4d-3c6f05546843", "external-id": "cl2-zone-554", "segmentation_id": 554, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a8ab9d1-00", "ovs_interfaceid": "9a8ab9d1-003d-492b-9a0b-726c69f6df9d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.864889] env[63202]: DEBUG oslo_vmware.api [None req-5423f696-c38b-4f13-988f-bd3e2591037c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385475, 'name': CloneVM_Task} progress is 94%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.879896] env[63202]: DEBUG nova.compute.manager [req-6b752bf3-42ae-4ff3-9871-cbadc1af43b6 req-5d181398-fc18-4ea9-a130-425871ef3de3 service nova] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Received event network-vif-plugged-f4c43c2b-ec12-47f7-84d8-4190afbf1b7d {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 921.880213] env[63202]: DEBUG oslo_concurrency.lockutils [req-6b752bf3-42ae-4ff3-9871-cbadc1af43b6 req-5d181398-fc18-4ea9-a130-425871ef3de3 service nova] Acquiring lock "af2bedc2-28ee-4679-ae38-1cceb2af05d6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.880447] env[63202]: DEBUG oslo_concurrency.lockutils [req-6b752bf3-42ae-4ff3-9871-cbadc1af43b6 req-5d181398-fc18-4ea9-a130-425871ef3de3 service nova] Lock "af2bedc2-28ee-4679-ae38-1cceb2af05d6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.880627] env[63202]: DEBUG oslo_concurrency.lockutils [req-6b752bf3-42ae-4ff3-9871-cbadc1af43b6 req-5d181398-fc18-4ea9-a130-425871ef3de3 service nova] Lock "af2bedc2-28ee-4679-ae38-1cceb2af05d6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.880806] env[63202]: DEBUG nova.compute.manager [req-6b752bf3-42ae-4ff3-9871-cbadc1af43b6 req-5d181398-fc18-4ea9-a130-425871ef3de3 service nova] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] No waiting events found dispatching network-vif-plugged-f4c43c2b-ec12-47f7-84d8-4190afbf1b7d {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 921.880970] env[63202]: WARNING nova.compute.manager [req-6b752bf3-42ae-4ff3-9871-cbadc1af43b6 req-5d181398-fc18-4ea9-a130-425871ef3de3 service nova] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Received unexpected event network-vif-plugged-f4c43c2b-ec12-47f7-84d8-4190afbf1b7d for instance with vm_state building and task_state spawning. [ 921.918348] env[63202]: DEBUG nova.compute.utils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 921.922137] env[63202]: DEBUG nova.compute.manager [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 921.922328] env[63202]: DEBUG nova.network.neutron [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 921.994773] env[63202]: DEBUG nova.policy [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9c4413afc60d4bf6aa0dd3a08b588b66', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'af1232a17bfd4bd3a3fe860f7da83994', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 922.032686] env[63202]: DEBUG nova.network.neutron [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Successfully updated port: f4c43c2b-ec12-47f7-84d8-4190afbf1b7d {{(pid=63202) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 922.081694] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-12584d87-b37a-41e6-b8aa-fb8129924ec5 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 922.082230] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cda90d17-37ad-4726-aae0-933e9ec4769a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.090348] env[63202]: DEBUG oslo_vmware.api [None req-12584d87-b37a-41e6-b8aa-fb8129924ec5 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Waiting for the task: (returnval){ [ 922.090348] env[63202]: value = "task-1385476" [ 922.090348] env[63202]: _type = "Task" [ 922.090348] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.103281] env[63202]: DEBUG oslo_vmware.api [None req-12584d87-b37a-41e6-b8aa-fb8129924ec5 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385476, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.301243] env[63202]: DEBUG nova.compute.manager [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 922.330260] env[63202]: DEBUG oslo_concurrency.lockutils [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Releasing lock "refresh_cache-753d190b-f4a4-4438-bc98-94564e3ec73d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.331989] env[63202]: DEBUG nova.compute.manager [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Instance network_info: |[{"id": "9a8ab9d1-003d-492b-9a0b-726c69f6df9d", "address": "fa:16:3e:62:14:16", "network": {"id": "ee6a8ee1-64c0-49e0-a3cb-d4df94f7c6e8", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1816904884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "90f7aa97c6a84ee2bd79518f6e7e2ad3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b107fab-ee71-47db-ad4d-3c6f05546843", "external-id": "cl2-zone-554", "segmentation_id": 554, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a8ab9d1-00", "ovs_interfaceid": "9a8ab9d1-003d-492b-9a0b-726c69f6df9d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 922.332140] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:62:14:16', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3b107fab-ee71-47db-ad4d-3c6f05546843', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9a8ab9d1-003d-492b-9a0b-726c69f6df9d', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 922.346915] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Creating folder: Project (90f7aa97c6a84ee2bd79518f6e7e2ad3). Parent ref: group-v294090. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 922.347313] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a2491a0f-b6fd-4182-9221-4bd841b18d41 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.363618] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Created folder: Project (90f7aa97c6a84ee2bd79518f6e7e2ad3) in parent group-v294090. [ 922.363952] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Creating folder: Instances. Parent ref: group-v294143. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 922.364644] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b6737304-9a73-49b3-bd15-1b25d4766dd7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.372608] env[63202]: DEBUG oslo_vmware.api [None req-5423f696-c38b-4f13-988f-bd3e2591037c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385475, 'name': CloneVM_Task} progress is 95%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.378923] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Created folder: Instances in parent group-v294143. [ 922.379197] env[63202]: DEBUG oslo.service.loopingcall [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 922.379395] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 922.379629] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9fc215fb-9051-48db-a9b9-78948862e057 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.401041] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 922.401041] env[63202]: value = "task-1385479" [ 922.401041] env[63202]: _type = "Task" [ 922.401041] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.409868] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385479, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.422977] env[63202]: DEBUG nova.compute.manager [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 922.507112] env[63202]: DEBUG nova.network.neutron [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Successfully created port: 718f088e-508c-4da3-97ec-11c84bfdab74 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 922.535534] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Acquiring lock "refresh_cache-af2bedc2-28ee-4679-ae38-1cceb2af05d6" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.535682] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Acquired lock "refresh_cache-af2bedc2-28ee-4679-ae38-1cceb2af05d6" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.535841] env[63202]: DEBUG nova.network.neutron [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 922.601727] env[63202]: DEBUG oslo_vmware.api [None req-12584d87-b37a-41e6-b8aa-fb8129924ec5 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385476, 'name': PowerOnVM_Task, 'duration_secs': 0.470947} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.604969] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-12584d87-b37a-41e6-b8aa-fb8129924ec5 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 922.605189] env[63202]: DEBUG nova.compute.manager [None req-12584d87-b37a-41e6-b8aa-fb8129924ec5 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 922.606517] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f59d7104-c1f5-4c22-89ef-8cee3e8e4f94 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.762700] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fb20ec5-c1d6-4b60-923c-b722447844c0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.772150] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a4573e9-7a12-4a47-a517-ca8d4d45d1fc {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.809624] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8048347f-d1f4-40f8-9f1d-526579f00167 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.820222] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ebb7130-dffd-4f31-8d13-7d10cd1d27f3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.826892] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.837323] env[63202]: DEBUG nova.compute.provider_tree [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 922.865123] env[63202]: DEBUG oslo_vmware.api [None req-5423f696-c38b-4f13-988f-bd3e2591037c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385475, 'name': CloneVM_Task, 'duration_secs': 1.155147} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.865416] env[63202]: INFO nova.virt.vmwareapi.vmops [None req-5423f696-c38b-4f13-988f-bd3e2591037c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Created linked-clone VM from snapshot [ 922.866186] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51fcc135-ae07-4eae-808c-e399d14783e4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.873458] env[63202]: DEBUG nova.virt.vmwareapi.images [None req-5423f696-c38b-4f13-988f-bd3e2591037c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Uploading image 376aee6c-1f98-4a60-858b-40c55f65b90e {{(pid=63202) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 922.894645] env[63202]: DEBUG oslo_vmware.rw_handles [None req-5423f696-c38b-4f13-988f-bd3e2591037c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 922.894645] env[63202]: value = "vm-294142" [ 922.894645] env[63202]: _type = "VirtualMachine" [ 922.894645] env[63202]: }. {{(pid=63202) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 922.894955] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-d4cce5ac-8a31-4dff-8fa4-bc96ea7f61c6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.902930] env[63202]: DEBUG oslo_vmware.rw_handles [None req-5423f696-c38b-4f13-988f-bd3e2591037c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lease: (returnval){ [ 922.902930] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]5206ceba-420c-10cc-f55f-f905a4c09cd4" [ 922.902930] env[63202]: _type = "HttpNfcLease" [ 922.902930] env[63202]: } obtained for exporting VM: (result){ [ 922.902930] env[63202]: value = "vm-294142" [ 922.902930] env[63202]: _type = "VirtualMachine" [ 922.902930] env[63202]: }. {{(pid=63202) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 922.903297] env[63202]: DEBUG oslo_vmware.api [None req-5423f696-c38b-4f13-988f-bd3e2591037c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for the lease: (returnval){ [ 922.903297] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]5206ceba-420c-10cc-f55f-f905a4c09cd4" [ 922.903297] env[63202]: _type = "HttpNfcLease" [ 922.903297] env[63202]: } to be ready. {{(pid=63202) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 922.915783] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385479, 'name': CreateVM_Task} progress is 25%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.917347] env[63202]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 922.917347] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]5206ceba-420c-10cc-f55f-f905a4c09cd4" [ 922.917347] env[63202]: _type = "HttpNfcLease" [ 922.917347] env[63202]: } is initializing. {{(pid=63202) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 923.070452] env[63202]: DEBUG nova.network.neutron [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 923.204809] env[63202]: DEBUG nova.network.neutron [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Updating instance_info_cache with network_info: [{"id": "f4c43c2b-ec12-47f7-84d8-4190afbf1b7d", "address": "fa:16:3e:5c:c9:fa", "network": {"id": "461615d3-d19d-4447-b32d-83c6e85b567d", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-2033752437-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "af1232a17bfd4bd3a3fe860f7da83994", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbf3349e-d05e-4d44-a011-c4b6e41af988", "external-id": "nsx-vlan-transportzone-196", "segmentation_id": 196, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf4c43c2b-ec", "ovs_interfaceid": "f4c43c2b-ec12-47f7-84d8-4190afbf1b7d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.340749] env[63202]: DEBUG nova.scheduler.client.report [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 923.413977] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385479, 'name': CreateVM_Task, 'duration_secs': 0.808415} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.414143] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 923.414817] env[63202]: DEBUG oslo_concurrency.lockutils [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.415353] env[63202]: DEBUG oslo_concurrency.lockutils [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.415353] env[63202]: DEBUG oslo_concurrency.lockutils [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 923.416958] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e0d09829-b0c2-48eb-a747-4621ff6e265b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.418562] env[63202]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 923.418562] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]5206ceba-420c-10cc-f55f-f905a4c09cd4" [ 923.418562] env[63202]: _type = "HttpNfcLease" [ 923.418562] env[63202]: } is ready. {{(pid=63202) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 923.418818] env[63202]: DEBUG oslo_vmware.rw_handles [None req-5423f696-c38b-4f13-988f-bd3e2591037c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 923.418818] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]5206ceba-420c-10cc-f55f-f905a4c09cd4" [ 923.418818] env[63202]: _type = "HttpNfcLease" [ 923.418818] env[63202]: }. {{(pid=63202) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 923.419885] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2662157-4ac9-4117-8eac-19e885f1821d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.423657] env[63202]: DEBUG oslo_vmware.api [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Waiting for the task: (returnval){ [ 923.423657] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52b873d8-68f6-a592-bef7-22b3a1091af9" [ 923.423657] env[63202]: _type = "Task" [ 923.423657] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.431068] env[63202]: DEBUG oslo_vmware.rw_handles [None req-5423f696-c38b-4f13-988f-bd3e2591037c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5290666b-725c-7b64-53b6-c8b95c3f9a76/disk-0.vmdk from lease info. {{(pid=63202) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 923.431068] env[63202]: DEBUG oslo_vmware.rw_handles [None req-5423f696-c38b-4f13-988f-bd3e2591037c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5290666b-725c-7b64-53b6-c8b95c3f9a76/disk-0.vmdk for reading. {{(pid=63202) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 923.487998] env[63202]: DEBUG nova.compute.manager [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 923.495582] env[63202]: DEBUG oslo_vmware.api [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52b873d8-68f6-a592-bef7-22b3a1091af9, 'name': SearchDatastore_Task, 'duration_secs': 0.011088} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.495895] env[63202]: DEBUG oslo_concurrency.lockutils [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.496134] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 923.496371] env[63202]: DEBUG oslo_concurrency.lockutils [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.496514] env[63202]: DEBUG oslo_concurrency.lockutils [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.496687] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 923.497237] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-32469559-e01a-4a53-b540-f6a8a3ed4dc4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.506815] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 923.507153] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 923.510112] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f8ea082-1949-4620-adf6-14c9d9c5148a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.516134] env[63202]: DEBUG oslo_vmware.api [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Waiting for the task: (returnval){ [ 923.516134] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52e2320d-69da-913b-e720-7ffa9b711e06" [ 923.516134] env[63202]: _type = "Task" [ 923.516134] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.521503] env[63202]: DEBUG nova.virt.hardware [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 923.522241] env[63202]: DEBUG nova.virt.hardware [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 923.522241] env[63202]: DEBUG nova.virt.hardware [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 923.522241] env[63202]: DEBUG nova.virt.hardware [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 923.522241] env[63202]: DEBUG nova.virt.hardware [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 923.522464] env[63202]: DEBUG nova.virt.hardware [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 923.522577] env[63202]: DEBUG nova.virt.hardware [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 923.522730] env[63202]: DEBUG nova.virt.hardware [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 923.522890] env[63202]: DEBUG nova.virt.hardware [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 923.523113] env[63202]: DEBUG nova.virt.hardware [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 923.523232] env[63202]: DEBUG nova.virt.hardware [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 923.524035] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9804d8f5-ff74-44e0-b566-5027fe2fdac4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.529280] env[63202]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-fb8f756c-0c57-4a05-89e1-550b711d21e3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.537739] env[63202]: DEBUG oslo_vmware.api [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52e2320d-69da-913b-e720-7ffa9b711e06, 'name': SearchDatastore_Task, 'duration_secs': 0.009449} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.541655] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90a3636f-19e7-484a-b861-ea7fa9a95381 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.548816] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2ce81a8a-c870-4ad5-b2f6-6b582ad6c351 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.565452] env[63202]: DEBUG oslo_vmware.api [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Waiting for the task: (returnval){ [ 923.565452] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]523bb66d-a0e3-41e2-ab1f-98dd8a3fb9f6" [ 923.565452] env[63202]: _type = "Task" [ 923.565452] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.575372] env[63202]: DEBUG oslo_vmware.api [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]523bb66d-a0e3-41e2-ab1f-98dd8a3fb9f6, 'name': SearchDatastore_Task, 'duration_secs': 0.009214} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.581195] env[63202]: DEBUG oslo_concurrency.lockutils [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.581475] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] 753d190b-f4a4-4438-bc98-94564e3ec73d/753d190b-f4a4-4438-bc98-94564e3ec73d.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 923.581933] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c6be6788-09db-4d9c-a286-8af46ced080b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.588607] env[63202]: DEBUG oslo_vmware.api [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Waiting for the task: (returnval){ [ 923.588607] env[63202]: value = "task-1385481" [ 923.588607] env[63202]: _type = "Task" [ 923.588607] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.598369] env[63202]: DEBUG oslo_vmware.api [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Task: {'id': task-1385481, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.708397] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Releasing lock "refresh_cache-af2bedc2-28ee-4679-ae38-1cceb2af05d6" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.708784] env[63202]: DEBUG nova.compute.manager [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Instance network_info: |[{"id": "f4c43c2b-ec12-47f7-84d8-4190afbf1b7d", "address": "fa:16:3e:5c:c9:fa", "network": {"id": "461615d3-d19d-4447-b32d-83c6e85b567d", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-2033752437-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "af1232a17bfd4bd3a3fe860f7da83994", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbf3349e-d05e-4d44-a011-c4b6e41af988", "external-id": "nsx-vlan-transportzone-196", "segmentation_id": 196, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf4c43c2b-ec", "ovs_interfaceid": "f4c43c2b-ec12-47f7-84d8-4190afbf1b7d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 923.709174] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5c:c9:fa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cbf3349e-d05e-4d44-a011-c4b6e41af988', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f4c43c2b-ec12-47f7-84d8-4190afbf1b7d', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 923.716988] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Creating folder: Project (af1232a17bfd4bd3a3fe860f7da83994). Parent ref: group-v294090. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 923.717325] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b6ce91e0-95fd-4aee-a5ee-0c7a318ad72b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.728084] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Created folder: Project (af1232a17bfd4bd3a3fe860f7da83994) in parent group-v294090. [ 923.728373] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Creating folder: Instances. Parent ref: group-v294146. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 923.728676] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-459f9e33-1282-4d15-b26e-39caa4447f44 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.737517] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Created folder: Instances in parent group-v294146. [ 923.737793] env[63202]: DEBUG oslo.service.loopingcall [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 923.739836] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 923.740284] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a087c0fc-fd6d-4df1-b446-8a0371e8d826 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.766060] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 923.766060] env[63202]: value = "task-1385484" [ 923.766060] env[63202]: _type = "Task" [ 923.766060] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.775949] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385484, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.845586] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.434s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.846801] env[63202]: DEBUG nova.compute.manager [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 923.849764] env[63202]: DEBUG oslo_concurrency.lockutils [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.512s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.851371] env[63202]: INFO nova.compute.claims [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 923.914396] env[63202]: DEBUG nova.compute.manager [req-197d0383-02be-41b7-852f-20befee5559c req-7dee2b58-e795-483a-bd0f-dd432034f385 service nova] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Received event network-changed-f4c43c2b-ec12-47f7-84d8-4190afbf1b7d {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 923.914554] env[63202]: DEBUG nova.compute.manager [req-197d0383-02be-41b7-852f-20befee5559c req-7dee2b58-e795-483a-bd0f-dd432034f385 service nova] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Refreshing instance network info cache due to event network-changed-f4c43c2b-ec12-47f7-84d8-4190afbf1b7d. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 923.914799] env[63202]: DEBUG oslo_concurrency.lockutils [req-197d0383-02be-41b7-852f-20befee5559c req-7dee2b58-e795-483a-bd0f-dd432034f385 service nova] Acquiring lock "refresh_cache-af2bedc2-28ee-4679-ae38-1cceb2af05d6" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.914946] env[63202]: DEBUG oslo_concurrency.lockutils [req-197d0383-02be-41b7-852f-20befee5559c req-7dee2b58-e795-483a-bd0f-dd432034f385 service nova] Acquired lock "refresh_cache-af2bedc2-28ee-4679-ae38-1cceb2af05d6" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.915145] env[63202]: DEBUG nova.network.neutron [req-197d0383-02be-41b7-852f-20befee5559c req-7dee2b58-e795-483a-bd0f-dd432034f385 service nova] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Refreshing network info cache for port f4c43c2b-ec12-47f7-84d8-4190afbf1b7d {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 924.079456] env[63202]: DEBUG nova.network.neutron [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Successfully updated port: 718f088e-508c-4da3-97ec-11c84bfdab74 {{(pid=63202) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 924.101864] env[63202]: DEBUG oslo_vmware.api [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Task: {'id': task-1385481, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.283754] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385484, 'name': CreateVM_Task} progress is 99%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.356670] env[63202]: DEBUG nova.compute.utils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 924.358601] env[63202]: DEBUG nova.compute.manager [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 924.361212] env[63202]: DEBUG nova.network.neutron [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 924.442058] env[63202]: DEBUG nova.policy [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9c4413afc60d4bf6aa0dd3a08b588b66', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'af1232a17bfd4bd3a3fe860f7da83994', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 924.581413] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Acquiring lock "refresh_cache-b1bccea9-2d79-431a-8be0-0a5ab293542a" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.581653] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Acquired lock "refresh_cache-b1bccea9-2d79-431a-8be0-0a5ab293542a" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.582034] env[63202]: DEBUG nova.network.neutron [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 924.605699] env[63202]: DEBUG oslo_vmware.api [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Task: {'id': task-1385481, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.56525} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.606115] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] 753d190b-f4a4-4438-bc98-94564e3ec73d/753d190b-f4a4-4438-bc98-94564e3ec73d.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 924.606513] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 924.606930] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-75d88088-9d2a-4cff-ac74-83740c1346c7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.631568] env[63202]: DEBUG oslo_vmware.api [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Waiting for the task: (returnval){ [ 924.631568] env[63202]: value = "task-1385485" [ 924.631568] env[63202]: _type = "Task" [ 924.631568] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.641701] env[63202]: DEBUG oslo_vmware.api [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Task: {'id': task-1385485, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.783174] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385484, 'name': CreateVM_Task, 'duration_secs': 0.601966} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.783174] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 924.784143] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.784444] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.784853] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 924.785187] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8e75a1ea-eaec-4c2e-837b-8a9d2dc48ef4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.792146] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Waiting for the task: (returnval){ [ 924.792146] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]523597f9-a6c9-3ea9-bd53-12bedafe149f" [ 924.792146] env[63202]: _type = "Task" [ 924.792146] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.803330] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]523597f9-a6c9-3ea9-bd53-12bedafe149f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.804242] env[63202]: DEBUG nova.network.neutron [req-197d0383-02be-41b7-852f-20befee5559c req-7dee2b58-e795-483a-bd0f-dd432034f385 service nova] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Updated VIF entry in instance network info cache for port f4c43c2b-ec12-47f7-84d8-4190afbf1b7d. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 924.804618] env[63202]: DEBUG nova.network.neutron [req-197d0383-02be-41b7-852f-20befee5559c req-7dee2b58-e795-483a-bd0f-dd432034f385 service nova] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Updating instance_info_cache with network_info: [{"id": "f4c43c2b-ec12-47f7-84d8-4190afbf1b7d", "address": "fa:16:3e:5c:c9:fa", "network": {"id": "461615d3-d19d-4447-b32d-83c6e85b567d", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-2033752437-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "af1232a17bfd4bd3a3fe860f7da83994", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbf3349e-d05e-4d44-a011-c4b6e41af988", "external-id": "nsx-vlan-transportzone-196", "segmentation_id": 196, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf4c43c2b-ec", "ovs_interfaceid": "f4c43c2b-ec12-47f7-84d8-4190afbf1b7d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.872561] env[63202]: DEBUG nova.compute.manager [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 924.896543] env[63202]: DEBUG nova.network.neutron [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Successfully created port: 79b59533-e84a-4ace-a615-596c8734838c {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 925.128234] env[63202]: DEBUG nova.network.neutron [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 925.145849] env[63202]: DEBUG oslo_vmware.api [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Task: {'id': task-1385485, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.154989} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.146414] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 925.148340] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c42bf7df-f8a0-4047-b48d-c12db1a44520 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.180877] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] 753d190b-f4a4-4438-bc98-94564e3ec73d/753d190b-f4a4-4438-bc98-94564e3ec73d.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 925.184321] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b1957303-41f2-41ae-9649-f1af9c963a6a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.209366] env[63202]: DEBUG oslo_vmware.api [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Waiting for the task: (returnval){ [ 925.209366] env[63202]: value = "task-1385486" [ 925.209366] env[63202]: _type = "Task" [ 925.209366] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.217858] env[63202]: DEBUG oslo_vmware.api [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Task: {'id': task-1385486, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.257635] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b3c0c37-6826-481f-993c-9bca707b9871 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.264338] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f2e62cb-372d-4846-b76c-8a86b7c98442 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.302021] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-743df602-c766-470b-af0b-6753de1fd028 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.309657] env[63202]: DEBUG oslo_concurrency.lockutils [req-197d0383-02be-41b7-852f-20befee5559c req-7dee2b58-e795-483a-bd0f-dd432034f385 service nova] Releasing lock "refresh_cache-af2bedc2-28ee-4679-ae38-1cceb2af05d6" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.310317] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]523597f9-a6c9-3ea9-bd53-12bedafe149f, 'name': SearchDatastore_Task, 'duration_secs': 0.011059} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.313432] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.313558] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 925.314140] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.314140] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.314345] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 925.314457] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ee0013e4-e469-47be-8832-f74a7a2574cd {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.317218] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5b4bcfb-af2c-497f-bb7f-8c26e6bf9850 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.330555] env[63202]: DEBUG nova.compute.provider_tree [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 925.334196] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 925.334196] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 925.334291] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-80306dbc-f4cd-4d48-86d3-fef85c859e63 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.341028] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Waiting for the task: (returnval){ [ 925.341028] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52aeee7e-0cb5-a9fd-e866-2d663f23cb20" [ 925.341028] env[63202]: _type = "Task" [ 925.341028] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.343704] env[63202]: DEBUG nova.network.neutron [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Updating instance_info_cache with network_info: [{"id": "718f088e-508c-4da3-97ec-11c84bfdab74", "address": "fa:16:3e:cb:64:d0", "network": {"id": "461615d3-d19d-4447-b32d-83c6e85b567d", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-2033752437-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "af1232a17bfd4bd3a3fe860f7da83994", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbf3349e-d05e-4d44-a011-c4b6e41af988", "external-id": "nsx-vlan-transportzone-196", "segmentation_id": 196, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap718f088e-50", "ovs_interfaceid": "718f088e-508c-4da3-97ec-11c84bfdab74", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.351576] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52aeee7e-0cb5-a9fd-e866-2d663f23cb20, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.723371] env[63202]: DEBUG oslo_vmware.api [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Task: {'id': task-1385486, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.838162] env[63202]: DEBUG nova.scheduler.client.report [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 925.852888] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Releasing lock "refresh_cache-b1bccea9-2d79-431a-8be0-0a5ab293542a" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.853371] env[63202]: DEBUG nova.compute.manager [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Instance network_info: |[{"id": "718f088e-508c-4da3-97ec-11c84bfdab74", "address": "fa:16:3e:cb:64:d0", "network": {"id": "461615d3-d19d-4447-b32d-83c6e85b567d", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-2033752437-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "af1232a17bfd4bd3a3fe860f7da83994", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbf3349e-d05e-4d44-a011-c4b6e41af988", "external-id": "nsx-vlan-transportzone-196", "segmentation_id": 196, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap718f088e-50", "ovs_interfaceid": "718f088e-508c-4da3-97ec-11c84bfdab74", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 925.856352] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cb:64:d0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cbf3349e-d05e-4d44-a011-c4b6e41af988', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '718f088e-508c-4da3-97ec-11c84bfdab74', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 925.865188] env[63202]: DEBUG oslo.service.loopingcall [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 925.866293] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 925.866840] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c3632996-98d3-492c-a259-b7c69eb9fd85 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.889910] env[63202]: DEBUG nova.compute.manager [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 925.892787] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52aeee7e-0cb5-a9fd-e866-2d663f23cb20, 'name': SearchDatastore_Task, 'duration_secs': 0.018722} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.896354] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9814d3a2-9fa9-4c84-a39f-0b3a8d791e3b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.898903] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 925.898903] env[63202]: value = "task-1385487" [ 925.898903] env[63202]: _type = "Task" [ 925.898903] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.904670] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Waiting for the task: (returnval){ [ 925.904670] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]525ca181-28df-72eb-c224-c9834df1ccdf" [ 925.904670] env[63202]: _type = "Task" [ 925.904670] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.915876] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385487, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.921336] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]525ca181-28df-72eb-c224-c9834df1ccdf, 'name': SearchDatastore_Task, 'duration_secs': 0.011047} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.921336] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.921336] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] af2bedc2-28ee-4679-ae38-1cceb2af05d6/af2bedc2-28ee-4679-ae38-1cceb2af05d6.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 925.921336] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e9f27b98-4c01-400d-82ef-78c118068f29 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.930327] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Waiting for the task: (returnval){ [ 925.930327] env[63202]: value = "task-1385488" [ 925.930327] env[63202]: _type = "Task" [ 925.930327] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.932650] env[63202]: DEBUG nova.virt.hardware [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 925.932943] env[63202]: DEBUG nova.virt.hardware [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 925.933181] env[63202]: DEBUG nova.virt.hardware [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 925.933439] env[63202]: DEBUG nova.virt.hardware [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 925.933640] env[63202]: DEBUG nova.virt.hardware [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 925.933859] env[63202]: DEBUG nova.virt.hardware [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 925.934142] env[63202]: DEBUG nova.virt.hardware [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 925.934368] env[63202]: DEBUG nova.virt.hardware [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 925.934595] env[63202]: DEBUG nova.virt.hardware [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 925.934828] env[63202]: DEBUG nova.virt.hardware [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 925.935111] env[63202]: DEBUG nova.virt.hardware [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 925.936157] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-007cf597-890c-4e7f-a112-dfaee9855c2b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.950527] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1c07aa8-0c88-491d-a49a-022802963ae9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.955079] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385488, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.992855] env[63202]: DEBUG nova.compute.manager [req-de6e95e8-b0c2-4b6e-a22a-562a7b7aa1aa req-6da3ce08-d227-4f89-bfe0-09306e4bbba3 service nova] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Received event network-vif-plugged-718f088e-508c-4da3-97ec-11c84bfdab74 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 925.992975] env[63202]: DEBUG oslo_concurrency.lockutils [req-de6e95e8-b0c2-4b6e-a22a-562a7b7aa1aa req-6da3ce08-d227-4f89-bfe0-09306e4bbba3 service nova] Acquiring lock "b1bccea9-2d79-431a-8be0-0a5ab293542a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.993210] env[63202]: DEBUG oslo_concurrency.lockutils [req-de6e95e8-b0c2-4b6e-a22a-562a7b7aa1aa req-6da3ce08-d227-4f89-bfe0-09306e4bbba3 service nova] Lock "b1bccea9-2d79-431a-8be0-0a5ab293542a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.993480] env[63202]: DEBUG oslo_concurrency.lockutils [req-de6e95e8-b0c2-4b6e-a22a-562a7b7aa1aa req-6da3ce08-d227-4f89-bfe0-09306e4bbba3 service nova] Lock "b1bccea9-2d79-431a-8be0-0a5ab293542a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.993598] env[63202]: DEBUG nova.compute.manager [req-de6e95e8-b0c2-4b6e-a22a-562a7b7aa1aa req-6da3ce08-d227-4f89-bfe0-09306e4bbba3 service nova] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] No waiting events found dispatching network-vif-plugged-718f088e-508c-4da3-97ec-11c84bfdab74 {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 925.993764] env[63202]: WARNING nova.compute.manager [req-de6e95e8-b0c2-4b6e-a22a-562a7b7aa1aa req-6da3ce08-d227-4f89-bfe0-09306e4bbba3 service nova] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Received unexpected event network-vif-plugged-718f088e-508c-4da3-97ec-11c84bfdab74 for instance with vm_state building and task_state spawning. [ 925.993962] env[63202]: DEBUG nova.compute.manager [req-de6e95e8-b0c2-4b6e-a22a-562a7b7aa1aa req-6da3ce08-d227-4f89-bfe0-09306e4bbba3 service nova] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Received event network-changed-718f088e-508c-4da3-97ec-11c84bfdab74 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 925.994085] env[63202]: DEBUG nova.compute.manager [req-de6e95e8-b0c2-4b6e-a22a-562a7b7aa1aa req-6da3ce08-d227-4f89-bfe0-09306e4bbba3 service nova] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Refreshing instance network info cache due to event network-changed-718f088e-508c-4da3-97ec-11c84bfdab74. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 925.994272] env[63202]: DEBUG oslo_concurrency.lockutils [req-de6e95e8-b0c2-4b6e-a22a-562a7b7aa1aa req-6da3ce08-d227-4f89-bfe0-09306e4bbba3 service nova] Acquiring lock "refresh_cache-b1bccea9-2d79-431a-8be0-0a5ab293542a" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.994407] env[63202]: DEBUG oslo_concurrency.lockutils [req-de6e95e8-b0c2-4b6e-a22a-562a7b7aa1aa req-6da3ce08-d227-4f89-bfe0-09306e4bbba3 service nova] Acquired lock "refresh_cache-b1bccea9-2d79-431a-8be0-0a5ab293542a" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.994563] env[63202]: DEBUG nova.network.neutron [req-de6e95e8-b0c2-4b6e-a22a-562a7b7aa1aa req-6da3ce08-d227-4f89-bfe0-09306e4bbba3 service nova] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Refreshing network info cache for port 718f088e-508c-4da3-97ec-11c84bfdab74 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 926.061094] env[63202]: DEBUG oslo_concurrency.lockutils [None req-61daa1db-4e4a-4b19-a510-d2d715a52389 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Acquiring lock "0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.061412] env[63202]: DEBUG oslo_concurrency.lockutils [None req-61daa1db-4e4a-4b19-a510-d2d715a52389 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lock "0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.221397] env[63202]: DEBUG oslo_vmware.api [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Task: {'id': task-1385486, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.225089] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d993524b-f10f-4ca3-a419-4ae825cd0a96 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Acquiring lock "61a6528e-4e4e-49b7-be7b-dd30bd8023d8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.225385] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d993524b-f10f-4ca3-a419-4ae825cd0a96 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Lock "61a6528e-4e4e-49b7-be7b-dd30bd8023d8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.225625] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d993524b-f10f-4ca3-a419-4ae825cd0a96 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Acquiring lock "61a6528e-4e4e-49b7-be7b-dd30bd8023d8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.225864] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d993524b-f10f-4ca3-a419-4ae825cd0a96 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Lock "61a6528e-4e4e-49b7-be7b-dd30bd8023d8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.226095] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d993524b-f10f-4ca3-a419-4ae825cd0a96 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Lock "61a6528e-4e4e-49b7-be7b-dd30bd8023d8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.231709] env[63202]: INFO nova.compute.manager [None req-d993524b-f10f-4ca3-a419-4ae825cd0a96 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Terminating instance [ 926.234151] env[63202]: DEBUG nova.compute.manager [None req-d993524b-f10f-4ca3-a419-4ae825cd0a96 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 926.234410] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d993524b-f10f-4ca3-a419-4ae825cd0a96 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 926.235478] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74b0a5f5-dd93-44a8-a2a9-6d9f7f28ab98 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.243843] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-d993524b-f10f-4ca3-a419-4ae825cd0a96 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 926.244187] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b85a4627-475f-4b6a-9c81-5bb3da82ea22 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.250845] env[63202]: DEBUG oslo_vmware.api [None req-d993524b-f10f-4ca3-a419-4ae825cd0a96 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Waiting for the task: (returnval){ [ 926.250845] env[63202]: value = "task-1385489" [ 926.250845] env[63202]: _type = "Task" [ 926.250845] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.258736] env[63202]: DEBUG oslo_vmware.api [None req-d993524b-f10f-4ca3-a419-4ae825cd0a96 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385489, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.346078] env[63202]: DEBUG oslo_concurrency.lockutils [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.496s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.346731] env[63202]: DEBUG nova.compute.manager [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 926.349610] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f347bd6c-7136-4427-b9a2-12f6c6163431 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.691s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.349813] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f347bd6c-7136-4427-b9a2-12f6c6163431 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.351967] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.440s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.353498] env[63202]: INFO nova.compute.claims [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 926.385922] env[63202]: INFO nova.scheduler.client.report [None req-f347bd6c-7136-4427-b9a2-12f6c6163431 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Deleted allocations for instance 50854a4e-8e4e-431b-8220-01bf8906b1c0 [ 926.410166] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385487, 'name': CreateVM_Task} progress is 25%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.445187] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385488, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.565567] env[63202]: DEBUG nova.compute.utils [None req-61daa1db-4e4a-4b19-a510-d2d715a52389 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 926.570596] env[63202]: DEBUG nova.network.neutron [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Successfully updated port: 79b59533-e84a-4ace-a615-596c8734838c {{(pid=63202) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 926.720643] env[63202]: DEBUG oslo_vmware.api [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Task: {'id': task-1385486, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.721593] env[63202]: DEBUG nova.network.neutron [req-de6e95e8-b0c2-4b6e-a22a-562a7b7aa1aa req-6da3ce08-d227-4f89-bfe0-09306e4bbba3 service nova] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Updated VIF entry in instance network info cache for port 718f088e-508c-4da3-97ec-11c84bfdab74. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 926.721950] env[63202]: DEBUG nova.network.neutron [req-de6e95e8-b0c2-4b6e-a22a-562a7b7aa1aa req-6da3ce08-d227-4f89-bfe0-09306e4bbba3 service nova] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Updating instance_info_cache with network_info: [{"id": "718f088e-508c-4da3-97ec-11c84bfdab74", "address": "fa:16:3e:cb:64:d0", "network": {"id": "461615d3-d19d-4447-b32d-83c6e85b567d", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-2033752437-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "af1232a17bfd4bd3a3fe860f7da83994", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbf3349e-d05e-4d44-a011-c4b6e41af988", "external-id": "nsx-vlan-transportzone-196", "segmentation_id": 196, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap718f088e-50", "ovs_interfaceid": "718f088e-508c-4da3-97ec-11c84bfdab74", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.760839] env[63202]: DEBUG oslo_vmware.api [None req-d993524b-f10f-4ca3-a419-4ae825cd0a96 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385489, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.862145] env[63202]: DEBUG nova.compute.utils [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 926.870150] env[63202]: DEBUG nova.compute.manager [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Not allocating networking since 'none' was specified. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 926.897410] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f347bd6c-7136-4427-b9a2-12f6c6163431 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "50854a4e-8e4e-431b-8220-01bf8906b1c0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.249s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.912603] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385487, 'name': CreateVM_Task} progress is 25%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.946782] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385488, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.56671} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.947164] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] af2bedc2-28ee-4679-ae38-1cceb2af05d6/af2bedc2-28ee-4679-ae38-1cceb2af05d6.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 926.947395] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 926.947675] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5febc97f-1208-4724-8907-98e9c6c9bf64 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.954945] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Waiting for the task: (returnval){ [ 926.954945] env[63202]: value = "task-1385490" [ 926.954945] env[63202]: _type = "Task" [ 926.954945] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.963802] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385490, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.069242] env[63202]: DEBUG oslo_concurrency.lockutils [None req-61daa1db-4e4a-4b19-a510-d2d715a52389 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lock "0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.072775] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Acquiring lock "refresh_cache-f3652744-e072-4700-80d4-b9eca414c5cb" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.072947] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Acquired lock "refresh_cache-f3652744-e072-4700-80d4-b9eca414c5cb" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.073118] env[63202]: DEBUG nova.network.neutron [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 927.224550] env[63202]: DEBUG oslo_concurrency.lockutils [req-de6e95e8-b0c2-4b6e-a22a-562a7b7aa1aa req-6da3ce08-d227-4f89-bfe0-09306e4bbba3 service nova] Releasing lock "refresh_cache-b1bccea9-2d79-431a-8be0-0a5ab293542a" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.228826] env[63202]: DEBUG oslo_vmware.api [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Task: {'id': task-1385486, 'name': ReconfigVM_Task, 'duration_secs': 1.669185} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.229146] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Reconfigured VM instance instance-00000049 to attach disk [datastore1] 753d190b-f4a4-4438-bc98-94564e3ec73d/753d190b-f4a4-4438-bc98-94564e3ec73d.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 927.229834] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b99b793f-84a1-43ff-8158-55a0892607b8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.236151] env[63202]: DEBUG oslo_vmware.api [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Waiting for the task: (returnval){ [ 927.236151] env[63202]: value = "task-1385491" [ 927.236151] env[63202]: _type = "Task" [ 927.236151] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.244953] env[63202]: DEBUG oslo_vmware.api [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Task: {'id': task-1385491, 'name': Rename_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.260716] env[63202]: DEBUG oslo_vmware.api [None req-d993524b-f10f-4ca3-a419-4ae825cd0a96 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385489, 'name': PowerOffVM_Task, 'duration_secs': 0.812253} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.260965] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-d993524b-f10f-4ca3-a419-4ae825cd0a96 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 927.261210] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d993524b-f10f-4ca3-a419-4ae825cd0a96 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 927.261492] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d7d7b663-bcab-41d6-b31d-0bdc2d5bfa2d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.326612] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d993524b-f10f-4ca3-a419-4ae825cd0a96 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 927.326869] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d993524b-f10f-4ca3-a419-4ae825cd0a96 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Deleting contents of the VM from datastore datastore2 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 927.327142] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-d993524b-f10f-4ca3-a419-4ae825cd0a96 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Deleting the datastore file [datastore2] 61a6528e-4e4e-49b7-be7b-dd30bd8023d8 {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 927.327476] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-718259bf-809b-4230-aa85-42cd7d99b773 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.334536] env[63202]: DEBUG oslo_vmware.api [None req-d993524b-f10f-4ca3-a419-4ae825cd0a96 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Waiting for the task: (returnval){ [ 927.334536] env[63202]: value = "task-1385493" [ 927.334536] env[63202]: _type = "Task" [ 927.334536] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.343216] env[63202]: DEBUG oslo_vmware.api [None req-d993524b-f10f-4ca3-a419-4ae825cd0a96 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385493, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.371251] env[63202]: DEBUG nova.compute.manager [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 927.413035] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385487, 'name': CreateVM_Task, 'duration_secs': 1.363146} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.416195] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 927.417096] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.417305] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.417645] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 927.417980] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f024dd6-077a-49b7-9770-552681eadd52 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.424024] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Waiting for the task: (returnval){ [ 927.424024] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52a9a170-d0cd-9960-d18f-0bbe01dd3419" [ 927.424024] env[63202]: _type = "Task" [ 927.424024] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.435808] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52a9a170-d0cd-9960-d18f-0bbe01dd3419, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.469471] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385490, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073187} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.469830] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 927.470947] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b380aab-27e5-4323-841b-bfca3b331b1f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.497748] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] af2bedc2-28ee-4679-ae38-1cceb2af05d6/af2bedc2-28ee-4679-ae38-1cceb2af05d6.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 927.501200] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-24174260-ad61-4cdc-8a8a-b2797f36849a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.520945] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Waiting for the task: (returnval){ [ 927.520945] env[63202]: value = "task-1385494" [ 927.520945] env[63202]: _type = "Task" [ 927.520945] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.529595] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385494, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.615406] env[63202]: DEBUG nova.network.neutron [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 927.669591] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f1f4223-329a-4795-8ae7-7d7aeb4eef7a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.676884] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4a8482e-7caa-4cf7-baef-3c773caba037 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.710920] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9504c39d-fb2e-43cd-995f-3874ea49ca4b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.718816] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a1a55c0-6971-4550-ba3b-a0475243810c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.734601] env[63202]: DEBUG nova.compute.provider_tree [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 927.745169] env[63202]: DEBUG oslo_vmware.api [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Task: {'id': task-1385491, 'name': Rename_Task, 'duration_secs': 0.230423} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.746074] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 927.746339] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-95747a26-6e7f-46b6-8f2c-04d2b492a85b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.754401] env[63202]: DEBUG oslo_vmware.api [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Waiting for the task: (returnval){ [ 927.754401] env[63202]: value = "task-1385495" [ 927.754401] env[63202]: _type = "Task" [ 927.754401] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.763115] env[63202]: DEBUG oslo_vmware.api [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Task: {'id': task-1385495, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.789865] env[63202]: DEBUG nova.network.neutron [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Updating instance_info_cache with network_info: [{"id": "79b59533-e84a-4ace-a615-596c8734838c", "address": "fa:16:3e:64:e2:2c", "network": {"id": "461615d3-d19d-4447-b32d-83c6e85b567d", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-2033752437-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "af1232a17bfd4bd3a3fe860f7da83994", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbf3349e-d05e-4d44-a011-c4b6e41af988", "external-id": "nsx-vlan-transportzone-196", "segmentation_id": 196, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79b59533-e8", "ovs_interfaceid": "79b59533-e84a-4ace-a615-596c8734838c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.844917] env[63202]: DEBUG oslo_vmware.api [None req-d993524b-f10f-4ca3-a419-4ae825cd0a96 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385493, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.166898} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.845713] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-d993524b-f10f-4ca3-a419-4ae825cd0a96 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 927.845713] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d993524b-f10f-4ca3-a419-4ae825cd0a96 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Deleted contents of the VM from datastore datastore2 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 927.845713] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d993524b-f10f-4ca3-a419-4ae825cd0a96 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 927.845713] env[63202]: INFO nova.compute.manager [None req-d993524b-f10f-4ca3-a419-4ae825cd0a96 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Took 1.61 seconds to destroy the instance on the hypervisor. [ 927.845928] env[63202]: DEBUG oslo.service.loopingcall [None req-d993524b-f10f-4ca3-a419-4ae825cd0a96 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 927.847041] env[63202]: DEBUG nova.compute.manager [-] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 927.847041] env[63202]: DEBUG nova.network.neutron [-] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 927.937864] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52a9a170-d0cd-9960-d18f-0bbe01dd3419, 'name': SearchDatastore_Task, 'duration_secs': 0.012972} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.937864] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.937864] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 927.937864] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.938021] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.938021] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 927.939400] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-918eef3c-dc2d-4d59-b398-6e51204c7b76 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.947845] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 927.948481] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 927.948953] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-028048d0-2505-4a43-ac4c-0b0597e37c4d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.955085] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Waiting for the task: (returnval){ [ 927.955085] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52ca0e7c-3dc1-f30d-7f9f-65293707b945" [ 927.955085] env[63202]: _type = "Task" [ 927.955085] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.963737] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52ca0e7c-3dc1-f30d-7f9f-65293707b945, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.025863] env[63202]: DEBUG nova.compute.manager [req-91435a36-e5b2-4d90-84d2-51a89e2f61dc req-cb139b04-1352-430c-8ca8-f2551b40bed6 service nova] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Received event network-vif-plugged-79b59533-e84a-4ace-a615-596c8734838c {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 928.025863] env[63202]: DEBUG oslo_concurrency.lockutils [req-91435a36-e5b2-4d90-84d2-51a89e2f61dc req-cb139b04-1352-430c-8ca8-f2551b40bed6 service nova] Acquiring lock "f3652744-e072-4700-80d4-b9eca414c5cb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.025863] env[63202]: DEBUG oslo_concurrency.lockutils [req-91435a36-e5b2-4d90-84d2-51a89e2f61dc req-cb139b04-1352-430c-8ca8-f2551b40bed6 service nova] Lock "f3652744-e072-4700-80d4-b9eca414c5cb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.025863] env[63202]: DEBUG oslo_concurrency.lockutils [req-91435a36-e5b2-4d90-84d2-51a89e2f61dc req-cb139b04-1352-430c-8ca8-f2551b40bed6 service nova] Lock "f3652744-e072-4700-80d4-b9eca414c5cb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.025863] env[63202]: DEBUG nova.compute.manager [req-91435a36-e5b2-4d90-84d2-51a89e2f61dc req-cb139b04-1352-430c-8ca8-f2551b40bed6 service nova] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] No waiting events found dispatching network-vif-plugged-79b59533-e84a-4ace-a615-596c8734838c {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 928.026721] env[63202]: WARNING nova.compute.manager [req-91435a36-e5b2-4d90-84d2-51a89e2f61dc req-cb139b04-1352-430c-8ca8-f2551b40bed6 service nova] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Received unexpected event network-vif-plugged-79b59533-e84a-4ace-a615-596c8734838c for instance with vm_state building and task_state spawning. [ 928.026721] env[63202]: DEBUG nova.compute.manager [req-91435a36-e5b2-4d90-84d2-51a89e2f61dc req-cb139b04-1352-430c-8ca8-f2551b40bed6 service nova] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Received event network-changed-79b59533-e84a-4ace-a615-596c8734838c {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 928.026721] env[63202]: DEBUG nova.compute.manager [req-91435a36-e5b2-4d90-84d2-51a89e2f61dc req-cb139b04-1352-430c-8ca8-f2551b40bed6 service nova] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Refreshing instance network info cache due to event network-changed-79b59533-e84a-4ace-a615-596c8734838c. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 928.026813] env[63202]: DEBUG oslo_concurrency.lockutils [req-91435a36-e5b2-4d90-84d2-51a89e2f61dc req-cb139b04-1352-430c-8ca8-f2551b40bed6 service nova] Acquiring lock "refresh_cache-f3652744-e072-4700-80d4-b9eca414c5cb" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.036819] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385494, 'name': ReconfigVM_Task, 'duration_secs': 0.340949} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.037224] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Reconfigured VM instance instance-0000004a to attach disk [datastore1] af2bedc2-28ee-4679-ae38-1cceb2af05d6/af2bedc2-28ee-4679-ae38-1cceb2af05d6.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 928.037987] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-da9b2b55-761f-453f-95c5-2da7515982ef {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.044447] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Waiting for the task: (returnval){ [ 928.044447] env[63202]: value = "task-1385496" [ 928.044447] env[63202]: _type = "Task" [ 928.044447] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.053579] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385496, 'name': Rename_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.147483] env[63202]: DEBUG oslo_concurrency.lockutils [None req-61daa1db-4e4a-4b19-a510-d2d715a52389 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Acquiring lock "0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.147795] env[63202]: DEBUG oslo_concurrency.lockutils [None req-61daa1db-4e4a-4b19-a510-d2d715a52389 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lock "0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.148103] env[63202]: INFO nova.compute.manager [None req-61daa1db-4e4a-4b19-a510-d2d715a52389 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Attaching volume 06e1ab70-2c2e-4848-9e08-4e17242e47b5 to /dev/sdb [ 928.191931] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e7ae6dd-d230-4412-bf85-d44015b6d45e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.199013] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff029842-0780-487f-829f-c8b13bf88466 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.214281] env[63202]: DEBUG nova.virt.block_device [None req-61daa1db-4e4a-4b19-a510-d2d715a52389 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Updating existing volume attachment record: 4130e3c1-da1a-49cb-a90c-229469f43fda {{(pid=63202) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 928.238587] env[63202]: DEBUG nova.scheduler.client.report [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 928.270059] env[63202]: DEBUG oslo_vmware.api [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Task: {'id': task-1385495, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.292736] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Releasing lock "refresh_cache-f3652744-e072-4700-80d4-b9eca414c5cb" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.293103] env[63202]: DEBUG nova.compute.manager [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Instance network_info: |[{"id": "79b59533-e84a-4ace-a615-596c8734838c", "address": "fa:16:3e:64:e2:2c", "network": {"id": "461615d3-d19d-4447-b32d-83c6e85b567d", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-2033752437-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "af1232a17bfd4bd3a3fe860f7da83994", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbf3349e-d05e-4d44-a011-c4b6e41af988", "external-id": "nsx-vlan-transportzone-196", "segmentation_id": 196, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79b59533-e8", "ovs_interfaceid": "79b59533-e84a-4ace-a615-596c8734838c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 928.293736] env[63202]: DEBUG oslo_concurrency.lockutils [req-91435a36-e5b2-4d90-84d2-51a89e2f61dc req-cb139b04-1352-430c-8ca8-f2551b40bed6 service nova] Acquired lock "refresh_cache-f3652744-e072-4700-80d4-b9eca414c5cb" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.293932] env[63202]: DEBUG nova.network.neutron [req-91435a36-e5b2-4d90-84d2-51a89e2f61dc req-cb139b04-1352-430c-8ca8-f2551b40bed6 service nova] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Refreshing network info cache for port 79b59533-e84a-4ace-a615-596c8734838c {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 928.295168] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:64:e2:2c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cbf3349e-d05e-4d44-a011-c4b6e41af988', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '79b59533-e84a-4ace-a615-596c8734838c', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 928.304938] env[63202]: DEBUG oslo.service.loopingcall [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 928.307043] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 928.307043] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1241c542-d54b-4536-8cfc-6b23a1d23a67 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.327388] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 928.327388] env[63202]: value = "task-1385497" [ 928.327388] env[63202]: _type = "Task" [ 928.327388] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.335967] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385497, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.383704] env[63202]: DEBUG nova.compute.manager [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 928.413498] env[63202]: DEBUG nova.virt.hardware [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 928.413792] env[63202]: DEBUG nova.virt.hardware [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 928.413971] env[63202]: DEBUG nova.virt.hardware [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 928.414191] env[63202]: DEBUG nova.virt.hardware [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 928.414389] env[63202]: DEBUG nova.virt.hardware [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 928.414505] env[63202]: DEBUG nova.virt.hardware [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 928.414737] env[63202]: DEBUG nova.virt.hardware [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 928.414911] env[63202]: DEBUG nova.virt.hardware [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 928.415109] env[63202]: DEBUG nova.virt.hardware [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 928.415278] env[63202]: DEBUG nova.virt.hardware [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 928.415451] env[63202]: DEBUG nova.virt.hardware [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 928.416462] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4b74046-9fcc-4ccf-a627-5075d86dc0fd {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.425307] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a52c0e8-40ee-4030-b53d-5368cf0780bb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.440757] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Instance VIF info [] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 928.447276] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Creating folder: Project (c1f94abed4cc4d85b6a5017f380518c4). Parent ref: group-v294090. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 928.447773] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-183e6199-5887-4603-acbb-278d04967610 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.462666] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Created folder: Project (c1f94abed4cc4d85b6a5017f380518c4) in parent group-v294090. [ 928.462915] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Creating folder: Instances. Parent ref: group-v294151. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 928.463296] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c139f8f3-bf5a-4863-8311-393f02cbfb08 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.470775] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52ca0e7c-3dc1-f30d-7f9f-65293707b945, 'name': SearchDatastore_Task, 'duration_secs': 0.010456} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.471774] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c3c42c3-8c58-4152-9a3b-fc8fde716e71 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.475411] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Created folder: Instances in parent group-v294151. [ 928.479337] env[63202]: DEBUG oslo.service.loopingcall [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 928.479337] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 928.479337] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-12a1fa81-b309-4c3c-b80a-738f514d3cb1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.490287] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Waiting for the task: (returnval){ [ 928.490287] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52d18c67-2c90-a047-1500-fc6e03839979" [ 928.490287] env[63202]: _type = "Task" [ 928.490287] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.495261] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 928.495261] env[63202]: value = "task-1385502" [ 928.495261] env[63202]: _type = "Task" [ 928.495261] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.502838] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52d18c67-2c90-a047-1500-fc6e03839979, 'name': SearchDatastore_Task, 'duration_secs': 0.013253} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.506246] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.506541] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] b1bccea9-2d79-431a-8be0-0a5ab293542a/b1bccea9-2d79-431a-8be0-0a5ab293542a.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 928.506963] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385502, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.507144] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-20069ee4-fc85-4968-a010-98d305a671b3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.515405] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Waiting for the task: (returnval){ [ 928.515405] env[63202]: value = "task-1385503" [ 928.515405] env[63202]: _type = "Task" [ 928.515405] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.530446] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385503, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.554143] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385496, 'name': Rename_Task, 'duration_secs': 0.169476} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.554542] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 928.554831] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-23b1b27a-ee3f-4e57-85b0-05feba4213c2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.560945] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Waiting for the task: (returnval){ [ 928.560945] env[63202]: value = "task-1385505" [ 928.560945] env[63202]: _type = "Task" [ 928.560945] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.571226] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385505, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.606065] env[63202]: DEBUG nova.network.neutron [-] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.645581] env[63202]: DEBUG oslo_concurrency.lockutils [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "2110dfa7-1795-4c8f-92bf-18921409c99f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.646022] env[63202]: DEBUG oslo_concurrency.lockutils [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "2110dfa7-1795-4c8f-92bf-18921409c99f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.744498] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.392s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.745069] env[63202]: DEBUG nova.compute.manager [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 928.748255] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.879s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.749900] env[63202]: INFO nova.compute.claims [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 928.766812] env[63202]: DEBUG oslo_vmware.api [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Task: {'id': task-1385495, 'name': PowerOnVM_Task, 'duration_secs': 0.748968} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.767102] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 928.767334] env[63202]: INFO nova.compute.manager [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Took 10.11 seconds to spawn the instance on the hypervisor. [ 928.767522] env[63202]: DEBUG nova.compute.manager [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 928.768534] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88e08510-fba5-4808-9cc1-b4326272388d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.836653] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385497, 'name': CreateVM_Task, 'duration_secs': 0.380015} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.836874] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 928.838749] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.838749] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.838749] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 928.838749] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9802be93-ee35-457c-90dc-dcd7953ba789 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.846657] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Waiting for the task: (returnval){ [ 928.846657] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52e07809-cf08-62db-d619-ac65ecef07b3" [ 928.846657] env[63202]: _type = "Task" [ 928.846657] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.857479] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52e07809-cf08-62db-d619-ac65ecef07b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.006428] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385502, 'name': CreateVM_Task, 'duration_secs': 0.361364} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.006599] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 929.007053] env[63202]: DEBUG oslo_concurrency.lockutils [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.007227] env[63202]: DEBUG oslo_concurrency.lockutils [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.007558] env[63202]: DEBUG oslo_concurrency.lockutils [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 929.007833] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3fb955a1-efbc-4848-b4ab-dfcc93ce6424 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.013871] env[63202]: DEBUG oslo_vmware.api [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Waiting for the task: (returnval){ [ 929.013871] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]5253437c-3a1e-3f87-937c-d1310aaf42d7" [ 929.013871] env[63202]: _type = "Task" [ 929.013871] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.025987] env[63202]: DEBUG oslo_vmware.api [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]5253437c-3a1e-3f87-937c-d1310aaf42d7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.029412] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385503, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.074037] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385505, 'name': PowerOnVM_Task} progress is 71%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.074037] env[63202]: DEBUG nova.network.neutron [req-91435a36-e5b2-4d90-84d2-51a89e2f61dc req-cb139b04-1352-430c-8ca8-f2551b40bed6 service nova] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Updated VIF entry in instance network info cache for port 79b59533-e84a-4ace-a615-596c8734838c. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 929.074381] env[63202]: DEBUG nova.network.neutron [req-91435a36-e5b2-4d90-84d2-51a89e2f61dc req-cb139b04-1352-430c-8ca8-f2551b40bed6 service nova] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Updating instance_info_cache with network_info: [{"id": "79b59533-e84a-4ace-a615-596c8734838c", "address": "fa:16:3e:64:e2:2c", "network": {"id": "461615d3-d19d-4447-b32d-83c6e85b567d", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-2033752437-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "af1232a17bfd4bd3a3fe860f7da83994", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbf3349e-d05e-4d44-a011-c4b6e41af988", "external-id": "nsx-vlan-transportzone-196", "segmentation_id": 196, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79b59533-e8", "ovs_interfaceid": "79b59533-e84a-4ace-a615-596c8734838c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.109091] env[63202]: INFO nova.compute.manager [-] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Took 1.26 seconds to deallocate network for instance. [ 929.255142] env[63202]: DEBUG nova.compute.utils [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 929.260245] env[63202]: DEBUG nova.compute.manager [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 929.260440] env[63202]: DEBUG nova.network.neutron [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 929.287036] env[63202]: INFO nova.compute.manager [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Took 34.65 seconds to build instance. [ 929.311926] env[63202]: DEBUG nova.policy [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6185cb1bba2f4200af4b143dd05425ad', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1ecf1ed8a25d4ee389e611648ed1c044', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 929.359552] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52e07809-cf08-62db-d619-ac65ecef07b3, 'name': SearchDatastore_Task, 'duration_secs': 0.059893} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.359924] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.360213] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 929.360471] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.360622] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.360800] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 929.361072] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0c1ba5ec-9230-464a-aea7-ea61ca6167f8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.370080] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 929.370305] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 929.371054] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-89f4c73c-0cc1-4273-8885-08073f903e49 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.377085] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Waiting for the task: (returnval){ [ 929.377085] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]5248942e-bb5e-ac37-6cca-7be832258d57" [ 929.377085] env[63202]: _type = "Task" [ 929.377085] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.385268] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]5248942e-bb5e-ac37-6cca-7be832258d57, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.531791] env[63202]: DEBUG oslo_vmware.api [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]5253437c-3a1e-3f87-937c-d1310aaf42d7, 'name': SearchDatastore_Task, 'duration_secs': 0.025131} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.536691] env[63202]: DEBUG oslo_concurrency.lockutils [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.536948] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 929.537203] env[63202]: DEBUG oslo_concurrency.lockutils [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.537357] env[63202]: DEBUG oslo_concurrency.lockutils [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.537537] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 929.537863] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385503, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.654412} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.538112] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0d36f600-2340-44f0-90fa-d60d8737438c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.540102] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] b1bccea9-2d79-431a-8be0-0a5ab293542a/b1bccea9-2d79-431a-8be0-0a5ab293542a.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 929.540357] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 929.540629] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a352dab5-2fe6-4642-9f54-141b89ff5b26 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.548081] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Waiting for the task: (returnval){ [ 929.548081] env[63202]: value = "task-1385506" [ 929.548081] env[63202]: _type = "Task" [ 929.548081] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.553172] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 929.553744] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 929.554596] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-10618874-89d4-4097-b138-c9a21e1ad1a9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.561431] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385506, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.568464] env[63202]: DEBUG oslo_vmware.api [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Waiting for the task: (returnval){ [ 929.568464] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52b1d3d5-2176-7c3d-705b-63f2325f97b3" [ 929.568464] env[63202]: _type = "Task" [ 929.568464] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.575914] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385505, 'name': PowerOnVM_Task, 'duration_secs': 0.881772} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.576627] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 929.576839] env[63202]: INFO nova.compute.manager [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Took 8.51 seconds to spawn the instance on the hypervisor. [ 929.577056] env[63202]: DEBUG nova.compute.manager [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 929.577978] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34621c38-42eb-491b-8ad8-16c0c80dc223 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.584835] env[63202]: DEBUG oslo_concurrency.lockutils [req-91435a36-e5b2-4d90-84d2-51a89e2f61dc req-cb139b04-1352-430c-8ca8-f2551b40bed6 service nova] Releasing lock "refresh_cache-f3652744-e072-4700-80d4-b9eca414c5cb" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.584835] env[63202]: DEBUG oslo_vmware.api [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52b1d3d5-2176-7c3d-705b-63f2325f97b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.599148] env[63202]: DEBUG nova.network.neutron [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Successfully created port: 41ce38c0-e94e-4de9-9504-0aaf673e3d03 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 929.616426] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d993524b-f10f-4ca3-a419-4ae825cd0a96 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.760712] env[63202]: DEBUG nova.compute.manager [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 929.788687] env[63202]: DEBUG oslo_concurrency.lockutils [None req-673e4529-7ede-4930-8957-94776fc44cb2 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Lock "753d190b-f4a4-4438-bc98-94564e3ec73d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 78.699s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.889201] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]5248942e-bb5e-ac37-6cca-7be832258d57, 'name': SearchDatastore_Task, 'duration_secs': 0.011978} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.892282] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1873eddf-a6e4-4627-b299-80c74d32e934 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.897386] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Waiting for the task: (returnval){ [ 929.897386] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]525c5fef-05b0-4d02-60e8-918e62187ce9" [ 929.897386] env[63202]: _type = "Task" [ 929.897386] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.906653] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]525c5fef-05b0-4d02-60e8-918e62187ce9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.037555] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2ce2d5b-89a8-4750-85a3-083e0b546e3b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.044798] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8582a36-6fa5-4a0e-b0d4-fa11b7736d78 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.051562] env[63202]: DEBUG nova.compute.manager [req-4357f91f-2d26-48e5-ac04-f85a60dada2e req-e51f34b2-55e1-4128-9eb4-22f00784e740 service nova] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Received event network-vif-deleted-38e58a7b-1f81-49a0-9b7d-25c507dc32e7 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 930.060066] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385506, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080568} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.083250] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 930.084854] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be239b8a-a9fa-45e8-9e53-bb81d88a4638 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.092224] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cee73d7-dd53-436b-91e5-fbf13548247b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.112249] env[63202]: DEBUG oslo_vmware.api [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52b1d3d5-2176-7c3d-705b-63f2325f97b3, 'name': SearchDatastore_Task, 'duration_secs': 0.015162} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.122345] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] b1bccea9-2d79-431a-8be0-0a5ab293542a/b1bccea9-2d79-431a-8be0-0a5ab293542a.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 930.124925] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5e712b61-d5ef-4b94-86aa-b3269c39d31d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.139824] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d004d419-97fd-41bc-a868-bf4dc28e0e0b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.144164] env[63202]: INFO nova.compute.manager [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Took 33.60 seconds to build instance. [ 930.145071] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3bba40c2-604d-41f7-80d5-578c8f90c597 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.163772] env[63202]: DEBUG oslo_vmware.api [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Waiting for the task: (returnval){ [ 930.163772] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]528bb69c-6093-c5b0-e82c-29724e639195" [ 930.163772] env[63202]: _type = "Task" [ 930.163772] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.164213] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Waiting for the task: (returnval){ [ 930.164213] env[63202]: value = "task-1385507" [ 930.164213] env[63202]: _type = "Task" [ 930.164213] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.164567] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3c8a2d3a-2d9f-4e04-858f-39e507432c13 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Acquiring lock "753d190b-f4a4-4438-bc98-94564e3ec73d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.164850] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3c8a2d3a-2d9f-4e04-858f-39e507432c13 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Lock "753d190b-f4a4-4438-bc98-94564e3ec73d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.165193] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3c8a2d3a-2d9f-4e04-858f-39e507432c13 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Acquiring lock "753d190b-f4a4-4438-bc98-94564e3ec73d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.165532] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3c8a2d3a-2d9f-4e04-858f-39e507432c13 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Lock "753d190b-f4a4-4438-bc98-94564e3ec73d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.165824] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3c8a2d3a-2d9f-4e04-858f-39e507432c13 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Lock "753d190b-f4a4-4438-bc98-94564e3ec73d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.168330] env[63202]: DEBUG nova.compute.provider_tree [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 930.170932] env[63202]: INFO nova.compute.manager [None req-3c8a2d3a-2d9f-4e04-858f-39e507432c13 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Terminating instance [ 930.173815] env[63202]: DEBUG nova.compute.manager [None req-3c8a2d3a-2d9f-4e04-858f-39e507432c13 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 930.174140] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-3c8a2d3a-2d9f-4e04-858f-39e507432c13 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 930.185037] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5732e2d7-2fdb-41e4-b759-c73f2eb6d77e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.193818] env[63202]: DEBUG oslo_vmware.api [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]528bb69c-6093-c5b0-e82c-29724e639195, 'name': SearchDatastore_Task, 'duration_secs': 0.029678} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.198654] env[63202]: DEBUG oslo_concurrency.lockutils [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.198955] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore2] fc7d3924-a624-4ea4-890c-7628595fb733/fc7d3924-a624-4ea4-890c-7628595fb733.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 930.199532] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385507, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.199817] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c8a2d3a-2d9f-4e04-858f-39e507432c13 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 930.199992] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-669c592b-f2d3-4a3b-9390-3256b4dccb22 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.201783] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8a4396e6-1271-46c4-ad39-26de1d6f8392 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.206692] env[63202]: DEBUG oslo_vmware.api [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Waiting for the task: (returnval){ [ 930.206692] env[63202]: value = "task-1385508" [ 930.206692] env[63202]: _type = "Task" [ 930.206692] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.207613] env[63202]: DEBUG oslo_vmware.api [None req-3c8a2d3a-2d9f-4e04-858f-39e507432c13 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Waiting for the task: (returnval){ [ 930.207613] env[63202]: value = "task-1385509" [ 930.207613] env[63202]: _type = "Task" [ 930.207613] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.218672] env[63202]: DEBUG oslo_vmware.api [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': task-1385508, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.221861] env[63202]: DEBUG oslo_vmware.api [None req-3c8a2d3a-2d9f-4e04-858f-39e507432c13 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Task: {'id': task-1385509, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.291715] env[63202]: DEBUG nova.compute.manager [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 930.409935] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]525c5fef-05b0-4d02-60e8-918e62187ce9, 'name': SearchDatastore_Task, 'duration_secs': 0.037252} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.410371] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.410652] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] f3652744-e072-4700-80d4-b9eca414c5cb/f3652744-e072-4700-80d4-b9eca414c5cb.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 930.410954] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cb5ea3d4-5e00-4413-af54-0897a7736d3c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.417930] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Waiting for the task: (returnval){ [ 930.417930] env[63202]: value = "task-1385510" [ 930.417930] env[63202]: _type = "Task" [ 930.417930] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.427025] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385510, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.648791] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Lock "af2bedc2-28ee-4679-ae38-1cceb2af05d6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 78.284s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.674578] env[63202]: DEBUG nova.scheduler.client.report [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 930.690744] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385507, 'name': ReconfigVM_Task, 'duration_secs': 0.485591} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.690937] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Reconfigured VM instance instance-0000004b to attach disk [datastore1] b1bccea9-2d79-431a-8be0-0a5ab293542a/b1bccea9-2d79-431a-8be0-0a5ab293542a.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 930.691734] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-223a52b4-dc5f-4341-967d-536681ad5617 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.699690] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Waiting for the task: (returnval){ [ 930.699690] env[63202]: value = "task-1385512" [ 930.699690] env[63202]: _type = "Task" [ 930.699690] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.710099] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385512, 'name': Rename_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.723165] env[63202]: DEBUG oslo_vmware.api [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': task-1385508, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.726763] env[63202]: DEBUG oslo_vmware.api [None req-3c8a2d3a-2d9f-4e04-858f-39e507432c13 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Task: {'id': task-1385509, 'name': PowerOffVM_Task, 'duration_secs': 0.216679} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.727079] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c8a2d3a-2d9f-4e04-858f-39e507432c13 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 930.727285] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-3c8a2d3a-2d9f-4e04-858f-39e507432c13 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 930.727595] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-47dc26a2-1d13-4be5-8c08-dbe9ff6494c5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.772863] env[63202]: DEBUG nova.compute.manager [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 930.799041] env[63202]: DEBUG nova.virt.hardware [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 930.799523] env[63202]: DEBUG nova.virt.hardware [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 930.799719] env[63202]: DEBUG nova.virt.hardware [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 930.799923] env[63202]: DEBUG nova.virt.hardware [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 930.800082] env[63202]: DEBUG nova.virt.hardware [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 930.800271] env[63202]: DEBUG nova.virt.hardware [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 930.800504] env[63202]: DEBUG nova.virt.hardware [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 930.800673] env[63202]: DEBUG nova.virt.hardware [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 930.800833] env[63202]: DEBUG nova.virt.hardware [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 930.801571] env[63202]: DEBUG nova.virt.hardware [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 930.801571] env[63202]: DEBUG nova.virt.hardware [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 930.806949] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac1a9d98-0ace-46d0-bfc0-394428016573 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.811426] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-3c8a2d3a-2d9f-4e04-858f-39e507432c13 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 930.811690] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-3c8a2d3a-2d9f-4e04-858f-39e507432c13 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Deleting contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 930.811881] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c8a2d3a-2d9f-4e04-858f-39e507432c13 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Deleting the datastore file [datastore1] 753d190b-f4a4-4438-bc98-94564e3ec73d {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 930.812599] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c1947c90-04ea-4a50-949b-b6dc758485ff {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.818493] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15ed9bbd-6aa5-4fe3-aea5-6a4a88166684 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.824009] env[63202]: DEBUG oslo_vmware.api [None req-3c8a2d3a-2d9f-4e04-858f-39e507432c13 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Waiting for the task: (returnval){ [ 930.824009] env[63202]: value = "task-1385514" [ 930.824009] env[63202]: _type = "Task" [ 930.824009] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.839476] env[63202]: DEBUG oslo_concurrency.lockutils [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.846606] env[63202]: DEBUG oslo_vmware.api [None req-3c8a2d3a-2d9f-4e04-858f-39e507432c13 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Task: {'id': task-1385514, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.928744] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385510, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.133415] env[63202]: DEBUG nova.network.neutron [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Successfully updated port: 41ce38c0-e94e-4de9-9504-0aaf673e3d03 {{(pid=63202) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 931.185691] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.437s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.186277] env[63202]: DEBUG nova.compute.manager [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 931.189161] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.506s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.190864] env[63202]: INFO nova.compute.claims [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 931.211811] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385512, 'name': Rename_Task, 'duration_secs': 0.188038} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.215860] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 931.216161] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a1188128-07db-4860-ad8f-8d8103868a3a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.223469] env[63202]: DEBUG oslo_vmware.api [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': task-1385508, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.752838} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.223850] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore2] fc7d3924-a624-4ea4-890c-7628595fb733/fc7d3924-a624-4ea4-890c-7628595fb733.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 931.223927] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 931.225416] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2f661457-03ab-4d60-b5fc-a947379c61c2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.227671] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Waiting for the task: (returnval){ [ 931.227671] env[63202]: value = "task-1385515" [ 931.227671] env[63202]: _type = "Task" [ 931.227671] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.234496] env[63202]: DEBUG oslo_vmware.api [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Waiting for the task: (returnval){ [ 931.234496] env[63202]: value = "task-1385516" [ 931.234496] env[63202]: _type = "Task" [ 931.234496] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.242346] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385515, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.248186] env[63202]: DEBUG oslo_vmware.api [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': task-1385516, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.334572] env[63202]: DEBUG oslo_vmware.api [None req-3c8a2d3a-2d9f-4e04-858f-39e507432c13 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Task: {'id': task-1385514, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.428332] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385510, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.834235} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.428653] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] f3652744-e072-4700-80d4-b9eca414c5cb/f3652744-e072-4700-80d4-b9eca414c5cb.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 931.429044] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 931.429325] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c51564e3-a2d9-4cea-a106-d2b3d136993a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.435163] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Waiting for the task: (returnval){ [ 931.435163] env[63202]: value = "task-1385517" [ 931.435163] env[63202]: _type = "Task" [ 931.435163] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.442984] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385517, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.635620] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Acquiring lock "refresh_cache-d0d6e380-9337-4f69-8434-6a33ac8a33d6" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.635815] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Acquired lock "refresh_cache-d0d6e380-9337-4f69-8434-6a33ac8a33d6" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.636790] env[63202]: DEBUG nova.network.neutron [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 931.695929] env[63202]: DEBUG nova.compute.utils [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 931.699339] env[63202]: DEBUG nova.compute.manager [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 931.699526] env[63202]: DEBUG nova.network.neutron [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 931.738906] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385515, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.740648] env[63202]: DEBUG nova.policy [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b59e7e62d92e4f5eb0fee41e8f5196d7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f508ec04edc844a19640a8a85f27e5b9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 931.750413] env[63202]: DEBUG oslo_vmware.api [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': task-1385516, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086243} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.750686] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 931.751494] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c091e33-07e1-49ec-8843-4d90ed27f749 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.771487] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] fc7d3924-a624-4ea4-890c-7628595fb733/fc7d3924-a624-4ea4-890c-7628595fb733.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 931.771719] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-134c0951-09c2-4f8d-af44-8a69bdbd54d2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.794299] env[63202]: DEBUG oslo_vmware.api [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Waiting for the task: (returnval){ [ 931.794299] env[63202]: value = "task-1385518" [ 931.794299] env[63202]: _type = "Task" [ 931.794299] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.802789] env[63202]: DEBUG oslo_vmware.api [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': task-1385518, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.835401] env[63202]: DEBUG oslo_vmware.api [None req-3c8a2d3a-2d9f-4e04-858f-39e507432c13 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Task: {'id': task-1385514, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.562745} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.835401] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c8a2d3a-2d9f-4e04-858f-39e507432c13 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 931.835401] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-3c8a2d3a-2d9f-4e04-858f-39e507432c13 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Deleted contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 931.835401] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-3c8a2d3a-2d9f-4e04-858f-39e507432c13 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 931.835401] env[63202]: INFO nova.compute.manager [None req-3c8a2d3a-2d9f-4e04-858f-39e507432c13 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Took 1.66 seconds to destroy the instance on the hypervisor. [ 931.835692] env[63202]: DEBUG oslo.service.loopingcall [None req-3c8a2d3a-2d9f-4e04-858f-39e507432c13 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 931.835692] env[63202]: DEBUG nova.compute.manager [-] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 931.835767] env[63202]: DEBUG nova.network.neutron [-] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 931.945084] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385517, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061031} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.947144] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 931.947968] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa54af12-afc6-455a-8917-97f4364092eb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.973424] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Reconfiguring VM instance instance-0000004c to attach disk [datastore1] f3652744-e072-4700-80d4-b9eca414c5cb/f3652744-e072-4700-80d4-b9eca414c5cb.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 931.973758] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-007d6159-5fad-4af5-b421-cedc5c9a25b5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.995357] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Waiting for the task: (returnval){ [ 931.995357] env[63202]: value = "task-1385519" [ 931.995357] env[63202]: _type = "Task" [ 931.995357] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.004679] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385519, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.084779] env[63202]: DEBUG nova.compute.manager [req-52a73856-fb54-4e83-aa17-2ae25a63751e req-5ad55bf3-0dce-404a-a873-ea0c3cba818f service nova] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Received event network-vif-plugged-41ce38c0-e94e-4de9-9504-0aaf673e3d03 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 932.085109] env[63202]: DEBUG oslo_concurrency.lockutils [req-52a73856-fb54-4e83-aa17-2ae25a63751e req-5ad55bf3-0dce-404a-a873-ea0c3cba818f service nova] Acquiring lock "d0d6e380-9337-4f69-8434-6a33ac8a33d6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.085367] env[63202]: DEBUG oslo_concurrency.lockutils [req-52a73856-fb54-4e83-aa17-2ae25a63751e req-5ad55bf3-0dce-404a-a873-ea0c3cba818f service nova] Lock "d0d6e380-9337-4f69-8434-6a33ac8a33d6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.085549] env[63202]: DEBUG oslo_concurrency.lockutils [req-52a73856-fb54-4e83-aa17-2ae25a63751e req-5ad55bf3-0dce-404a-a873-ea0c3cba818f service nova] Lock "d0d6e380-9337-4f69-8434-6a33ac8a33d6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.085739] env[63202]: DEBUG nova.compute.manager [req-52a73856-fb54-4e83-aa17-2ae25a63751e req-5ad55bf3-0dce-404a-a873-ea0c3cba818f service nova] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] No waiting events found dispatching network-vif-plugged-41ce38c0-e94e-4de9-9504-0aaf673e3d03 {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 932.085924] env[63202]: WARNING nova.compute.manager [req-52a73856-fb54-4e83-aa17-2ae25a63751e req-5ad55bf3-0dce-404a-a873-ea0c3cba818f service nova] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Received unexpected event network-vif-plugged-41ce38c0-e94e-4de9-9504-0aaf673e3d03 for instance with vm_state building and task_state spawning. [ 932.086146] env[63202]: DEBUG nova.compute.manager [req-52a73856-fb54-4e83-aa17-2ae25a63751e req-5ad55bf3-0dce-404a-a873-ea0c3cba818f service nova] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Received event network-changed-41ce38c0-e94e-4de9-9504-0aaf673e3d03 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 932.086344] env[63202]: DEBUG nova.compute.manager [req-52a73856-fb54-4e83-aa17-2ae25a63751e req-5ad55bf3-0dce-404a-a873-ea0c3cba818f service nova] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Refreshing instance network info cache due to event network-changed-41ce38c0-e94e-4de9-9504-0aaf673e3d03. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 932.086557] env[63202]: DEBUG oslo_concurrency.lockutils [req-52a73856-fb54-4e83-aa17-2ae25a63751e req-5ad55bf3-0dce-404a-a873-ea0c3cba818f service nova] Acquiring lock "refresh_cache-d0d6e380-9337-4f69-8434-6a33ac8a33d6" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.184706] env[63202]: DEBUG nova.network.neutron [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 932.187176] env[63202]: DEBUG nova.network.neutron [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Successfully created port: faa30f43-2e38-4dbb-97ca-1b4c17565310 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 932.200197] env[63202]: DEBUG nova.compute.manager [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 932.247236] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385515, 'name': PowerOnVM_Task, 'duration_secs': 0.786576} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.250728] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 932.250989] env[63202]: INFO nova.compute.manager [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Took 8.76 seconds to spawn the instance on the hypervisor. [ 932.251227] env[63202]: DEBUG nova.compute.manager [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 932.252737] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c50d8fa4-40d7-414f-bcd7-1c6ef7ae90d3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.306418] env[63202]: DEBUG oslo_vmware.api [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': task-1385518, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.509667] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385519, 'name': ReconfigVM_Task, 'duration_secs': 0.474025} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.509949] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Reconfigured VM instance instance-0000004c to attach disk [datastore1] f3652744-e072-4700-80d4-b9eca414c5cb/f3652744-e072-4700-80d4-b9eca414c5cb.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 932.510717] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-52f7f41d-35e0-446a-b1a5-2eb2d43543e9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.516832] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Waiting for the task: (returnval){ [ 932.516832] env[63202]: value = "task-1385520" [ 932.516832] env[63202]: _type = "Task" [ 932.516832] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.532265] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385520, 'name': Rename_Task} progress is 6%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.532666] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99801430-be29-45d8-b732-b3a42f05e334 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.535843] env[63202]: DEBUG nova.network.neutron [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Updating instance_info_cache with network_info: [{"id": "41ce38c0-e94e-4de9-9504-0aaf673e3d03", "address": "fa:16:3e:98:c1:c6", "network": {"id": "1e700e94-1088-429b-b630-4e99f7522ecd", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1316793813-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ecf1ed8a25d4ee389e611648ed1c044", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "57c65f87-60fd-4882-ab30-31db49131b46", "external-id": "nsx-vlan-transportzone-610", "segmentation_id": 610, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41ce38c0-e9", "ovs_interfaceid": "41ce38c0-e94e-4de9-9504-0aaf673e3d03", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.544360] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca2e374c-b2b0-487b-afc4-28b62ee3d582 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.575942] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4577357c-3f89-446e-99ff-6ae86ac47c2a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.584191] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cbee267-97eb-4993-b106-3327762a008c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.598464] env[63202]: DEBUG nova.compute.provider_tree [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 932.772740] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-61daa1db-4e4a-4b19-a510-d2d715a52389 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Volume attach. Driver type: vmdk {{(pid=63202) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 932.772980] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-61daa1db-4e4a-4b19-a510-d2d715a52389 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294155', 'volume_id': '06e1ab70-2c2e-4848-9e08-4e17242e47b5', 'name': 'volume-06e1ab70-2c2e-4848-9e08-4e17242e47b5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb', 'attached_at': '', 'detached_at': '', 'volume_id': '06e1ab70-2c2e-4848-9e08-4e17242e47b5', 'serial': '06e1ab70-2c2e-4848-9e08-4e17242e47b5'} {{(pid=63202) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 932.773990] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55364cce-cef2-476f-bf8e-df633b0623fe {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.782970] env[63202]: INFO nova.compute.manager [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Took 36.15 seconds to build instance. [ 932.799512] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66b8ad56-3d27-4c28-aa46-cd7dc0ae7b7f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.802884] env[63202]: DEBUG nova.network.neutron [-] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.809995] env[63202]: DEBUG oslo_vmware.api [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': task-1385518, 'name': ReconfigVM_Task, 'duration_secs': 0.554776} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.825721] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Reconfigured VM instance instance-0000004d to attach disk [datastore2] fc7d3924-a624-4ea4-890c-7628595fb733/fc7d3924-a624-4ea4-890c-7628595fb733.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 932.834991] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-61daa1db-4e4a-4b19-a510-d2d715a52389 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] volume-06e1ab70-2c2e-4848-9e08-4e17242e47b5/volume-06e1ab70-2c2e-4848-9e08-4e17242e47b5.vmdk or device None with type thin {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 932.835385] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-851426bb-c43f-491a-892a-e0352c6db0c7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.837718] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aa6e8b47-7dea-4656-95ec-a62d769d449a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.857709] env[63202]: DEBUG oslo_vmware.api [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Waiting for the task: (returnval){ [ 932.857709] env[63202]: value = "task-1385521" [ 932.857709] env[63202]: _type = "Task" [ 932.857709] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.858887] env[63202]: DEBUG oslo_vmware.api [None req-61daa1db-4e4a-4b19-a510-d2d715a52389 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Waiting for the task: (returnval){ [ 932.858887] env[63202]: value = "task-1385522" [ 932.858887] env[63202]: _type = "Task" [ 932.858887] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.870332] env[63202]: DEBUG oslo_vmware.api [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': task-1385521, 'name': Rename_Task} progress is 6%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.873638] env[63202]: DEBUG oslo_vmware.api [None req-61daa1db-4e4a-4b19-a510-d2d715a52389 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385522, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.029817] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385520, 'name': Rename_Task, 'duration_secs': 0.328585} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.030124] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 933.030488] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3edc6c72-1f40-4c6c-9738-16a3dfe9631a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.039024] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Releasing lock "refresh_cache-d0d6e380-9337-4f69-8434-6a33ac8a33d6" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.039024] env[63202]: DEBUG nova.compute.manager [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Instance network_info: |[{"id": "41ce38c0-e94e-4de9-9504-0aaf673e3d03", "address": "fa:16:3e:98:c1:c6", "network": {"id": "1e700e94-1088-429b-b630-4e99f7522ecd", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1316793813-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ecf1ed8a25d4ee389e611648ed1c044", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "57c65f87-60fd-4882-ab30-31db49131b46", "external-id": "nsx-vlan-transportzone-610", "segmentation_id": 610, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41ce38c0-e9", "ovs_interfaceid": "41ce38c0-e94e-4de9-9504-0aaf673e3d03", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 933.039233] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Waiting for the task: (returnval){ [ 933.039233] env[63202]: value = "task-1385523" [ 933.039233] env[63202]: _type = "Task" [ 933.039233] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.039233] env[63202]: DEBUG oslo_concurrency.lockutils [req-52a73856-fb54-4e83-aa17-2ae25a63751e req-5ad55bf3-0dce-404a-a873-ea0c3cba818f service nova] Acquired lock "refresh_cache-d0d6e380-9337-4f69-8434-6a33ac8a33d6" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.039233] env[63202]: DEBUG nova.network.neutron [req-52a73856-fb54-4e83-aa17-2ae25a63751e req-5ad55bf3-0dce-404a-a873-ea0c3cba818f service nova] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Refreshing network info cache for port 41ce38c0-e94e-4de9-9504-0aaf673e3d03 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 933.040389] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:98:c1:c6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '57c65f87-60fd-4882-ab30-31db49131b46', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '41ce38c0-e94e-4de9-9504-0aaf673e3d03', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 933.047892] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Creating folder: Project (1ecf1ed8a25d4ee389e611648ed1c044). Parent ref: group-v294090. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 933.051045] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-476e87f1-ed5a-4b02-ac1d-c49c245c1189 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.062090] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385523, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.064715] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Created folder: Project (1ecf1ed8a25d4ee389e611648ed1c044) in parent group-v294090. [ 933.064923] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Creating folder: Instances. Parent ref: group-v294156. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 933.065209] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-88896781-2697-4cf1-87da-2a557440444f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.073834] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Created folder: Instances in parent group-v294156. [ 933.074133] env[63202]: DEBUG oslo.service.loopingcall [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 933.074805] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 933.074805] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0b865bca-a376-47a4-9e29-2f9d89139f87 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.095803] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 933.095803] env[63202]: value = "task-1385526" [ 933.095803] env[63202]: _type = "Task" [ 933.095803] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.105015] env[63202]: DEBUG nova.scheduler.client.report [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 933.107968] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385526, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.215675] env[63202]: DEBUG nova.compute.manager [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 933.240793] env[63202]: DEBUG nova.virt.hardware [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 933.241316] env[63202]: DEBUG nova.virt.hardware [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 933.241572] env[63202]: DEBUG nova.virt.hardware [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 933.241885] env[63202]: DEBUG nova.virt.hardware [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 933.242155] env[63202]: DEBUG nova.virt.hardware [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 933.242363] env[63202]: DEBUG nova.virt.hardware [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 933.242670] env[63202]: DEBUG nova.virt.hardware [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 933.242912] env[63202]: DEBUG nova.virt.hardware [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 933.243230] env[63202]: DEBUG nova.virt.hardware [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 933.243490] env[63202]: DEBUG nova.virt.hardware [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 933.243749] env[63202]: DEBUG nova.virt.hardware [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 933.244929] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d126f666-28a8-4d09-90f5-b7c41d318658 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.255386] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e0860b8-46e0-4114-9a2e-04ee2ddcdb9b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.275155] env[63202]: DEBUG nova.network.neutron [req-52a73856-fb54-4e83-aa17-2ae25a63751e req-5ad55bf3-0dce-404a-a873-ea0c3cba818f service nova] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Updated VIF entry in instance network info cache for port 41ce38c0-e94e-4de9-9504-0aaf673e3d03. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 933.275581] env[63202]: DEBUG nova.network.neutron [req-52a73856-fb54-4e83-aa17-2ae25a63751e req-5ad55bf3-0dce-404a-a873-ea0c3cba818f service nova] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Updating instance_info_cache with network_info: [{"id": "41ce38c0-e94e-4de9-9504-0aaf673e3d03", "address": "fa:16:3e:98:c1:c6", "network": {"id": "1e700e94-1088-429b-b630-4e99f7522ecd", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1316793813-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ecf1ed8a25d4ee389e611648ed1c044", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "57c65f87-60fd-4882-ab30-31db49131b46", "external-id": "nsx-vlan-transportzone-610", "segmentation_id": 610, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41ce38c0-e9", "ovs_interfaceid": "41ce38c0-e94e-4de9-9504-0aaf673e3d03", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.296448] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Lock "b1bccea9-2d79-431a-8be0-0a5ab293542a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 80.908s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.305878] env[63202]: INFO nova.compute.manager [-] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Took 1.47 seconds to deallocate network for instance. [ 933.373546] env[63202]: DEBUG oslo_vmware.api [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': task-1385521, 'name': Rename_Task, 'duration_secs': 0.218685} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.377185] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 933.377604] env[63202]: DEBUG oslo_vmware.api [None req-61daa1db-4e4a-4b19-a510-d2d715a52389 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385522, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.377855] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bf856d93-014d-4792-9163-7cefe3394566 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.385173] env[63202]: DEBUG oslo_vmware.api [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Waiting for the task: (returnval){ [ 933.385173] env[63202]: value = "task-1385527" [ 933.385173] env[63202]: _type = "Task" [ 933.385173] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.396621] env[63202]: DEBUG oslo_vmware.api [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': task-1385527, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.558677] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385523, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.605719] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385526, 'name': CreateVM_Task} progress is 25%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.610019] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.421s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.610568] env[63202]: DEBUG nova.compute.manager [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 933.613673] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.266s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.615138] env[63202]: INFO nova.compute.claims [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 933.757921] env[63202]: DEBUG nova.network.neutron [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Successfully updated port: faa30f43-2e38-4dbb-97ca-1b4c17565310 {{(pid=63202) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 933.778885] env[63202]: DEBUG oslo_concurrency.lockutils [req-52a73856-fb54-4e83-aa17-2ae25a63751e req-5ad55bf3-0dce-404a-a873-ea0c3cba818f service nova] Releasing lock "refresh_cache-d0d6e380-9337-4f69-8434-6a33ac8a33d6" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.812370] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3c8a2d3a-2d9f-4e04-858f-39e507432c13 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.874260] env[63202]: DEBUG oslo_vmware.api [None req-61daa1db-4e4a-4b19-a510-d2d715a52389 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385522, 'name': ReconfigVM_Task, 'duration_secs': 0.581051} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.874569] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-61daa1db-4e4a-4b19-a510-d2d715a52389 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Reconfigured VM instance instance-0000003f to attach disk [datastore2] volume-06e1ab70-2c2e-4848-9e08-4e17242e47b5/volume-06e1ab70-2c2e-4848-9e08-4e17242e47b5.vmdk or device None with type thin {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 933.879582] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-72d8516f-b7fd-466c-a3b5-9c5a774d0e3e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.898682] env[63202]: DEBUG oslo_vmware.api [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': task-1385527, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.900388] env[63202]: DEBUG oslo_vmware.api [None req-61daa1db-4e4a-4b19-a510-d2d715a52389 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Waiting for the task: (returnval){ [ 933.900388] env[63202]: value = "task-1385528" [ 933.900388] env[63202]: _type = "Task" [ 933.900388] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.908558] env[63202]: DEBUG oslo_vmware.api [None req-61daa1db-4e4a-4b19-a510-d2d715a52389 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385528, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.058410] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385523, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.108308] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385526, 'name': CreateVM_Task} progress is 25%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.122207] env[63202]: DEBUG nova.compute.utils [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 934.123686] env[63202]: DEBUG nova.compute.manager [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 934.123848] env[63202]: DEBUG nova.network.neutron [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 934.128883] env[63202]: DEBUG nova.compute.manager [req-fbee6873-ecff-48c4-a2da-f6079c4d6446 req-3c6a4385-3a68-4304-85c2-c32a8e79f696 service nova] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Received event network-vif-deleted-9a8ab9d1-003d-492b-9a0b-726c69f6df9d {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 934.129150] env[63202]: DEBUG nova.compute.manager [req-fbee6873-ecff-48c4-a2da-f6079c4d6446 req-3c6a4385-3a68-4304-85c2-c32a8e79f696 service nova] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Received event network-vif-plugged-faa30f43-2e38-4dbb-97ca-1b4c17565310 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 934.129345] env[63202]: DEBUG oslo_concurrency.lockutils [req-fbee6873-ecff-48c4-a2da-f6079c4d6446 req-3c6a4385-3a68-4304-85c2-c32a8e79f696 service nova] Acquiring lock "8755bceb-d510-4429-bd98-d6a63faf739d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.129591] env[63202]: DEBUG oslo_concurrency.lockutils [req-fbee6873-ecff-48c4-a2da-f6079c4d6446 req-3c6a4385-3a68-4304-85c2-c32a8e79f696 service nova] Lock "8755bceb-d510-4429-bd98-d6a63faf739d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.129806] env[63202]: DEBUG oslo_concurrency.lockutils [req-fbee6873-ecff-48c4-a2da-f6079c4d6446 req-3c6a4385-3a68-4304-85c2-c32a8e79f696 service nova] Lock "8755bceb-d510-4429-bd98-d6a63faf739d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.129865] env[63202]: DEBUG nova.compute.manager [req-fbee6873-ecff-48c4-a2da-f6079c4d6446 req-3c6a4385-3a68-4304-85c2-c32a8e79f696 service nova] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] No waiting events found dispatching network-vif-plugged-faa30f43-2e38-4dbb-97ca-1b4c17565310 {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 934.130020] env[63202]: WARNING nova.compute.manager [req-fbee6873-ecff-48c4-a2da-f6079c4d6446 req-3c6a4385-3a68-4304-85c2-c32a8e79f696 service nova] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Received unexpected event network-vif-plugged-faa30f43-2e38-4dbb-97ca-1b4c17565310 for instance with vm_state building and task_state spawning. [ 934.130242] env[63202]: DEBUG nova.compute.manager [req-fbee6873-ecff-48c4-a2da-f6079c4d6446 req-3c6a4385-3a68-4304-85c2-c32a8e79f696 service nova] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Received event network-changed-faa30f43-2e38-4dbb-97ca-1b4c17565310 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 934.130414] env[63202]: DEBUG nova.compute.manager [req-fbee6873-ecff-48c4-a2da-f6079c4d6446 req-3c6a4385-3a68-4304-85c2-c32a8e79f696 service nova] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Refreshing instance network info cache due to event network-changed-faa30f43-2e38-4dbb-97ca-1b4c17565310. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 934.130598] env[63202]: DEBUG oslo_concurrency.lockutils [req-fbee6873-ecff-48c4-a2da-f6079c4d6446 req-3c6a4385-3a68-4304-85c2-c32a8e79f696 service nova] Acquiring lock "refresh_cache-8755bceb-d510-4429-bd98-d6a63faf739d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.131250] env[63202]: DEBUG oslo_concurrency.lockutils [req-fbee6873-ecff-48c4-a2da-f6079c4d6446 req-3c6a4385-3a68-4304-85c2-c32a8e79f696 service nova] Acquired lock "refresh_cache-8755bceb-d510-4429-bd98-d6a63faf739d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.131250] env[63202]: DEBUG nova.network.neutron [req-fbee6873-ecff-48c4-a2da-f6079c4d6446 req-3c6a4385-3a68-4304-85c2-c32a8e79f696 service nova] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Refreshing network info cache for port faa30f43-2e38-4dbb-97ca-1b4c17565310 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 934.179933] env[63202]: DEBUG nova.policy [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '185fb4e52283440c8cc1b226cb7c6c71', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '417f815aec0c451a8ccd713ad6c64ad2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 934.261087] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "refresh_cache-8755bceb-d510-4429-bd98-d6a63faf739d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.400899] env[63202]: DEBUG oslo_vmware.api [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': task-1385527, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.411022] env[63202]: DEBUG oslo_vmware.api [None req-61daa1db-4e4a-4b19-a510-d2d715a52389 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385528, 'name': ReconfigVM_Task, 'duration_secs': 0.142463} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.411022] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-61daa1db-4e4a-4b19-a510-d2d715a52389 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294155', 'volume_id': '06e1ab70-2c2e-4848-9e08-4e17242e47b5', 'name': 'volume-06e1ab70-2c2e-4848-9e08-4e17242e47b5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb', 'attached_at': '', 'detached_at': '', 'volume_id': '06e1ab70-2c2e-4848-9e08-4e17242e47b5', 'serial': '06e1ab70-2c2e-4848-9e08-4e17242e47b5'} {{(pid=63202) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 934.447944] env[63202]: DEBUG nova.network.neutron [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Successfully created port: be85354c-2bf7-4c06-898f-7061cdfada23 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 934.484630] env[63202]: DEBUG oslo_vmware.rw_handles [None req-5423f696-c38b-4f13-988f-bd3e2591037c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5290666b-725c-7b64-53b6-c8b95c3f9a76/disk-0.vmdk. {{(pid=63202) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 934.485628] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5a6bec8-31cf-428b-b658-777340179eaf {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.492420] env[63202]: DEBUG oslo_vmware.rw_handles [None req-5423f696-c38b-4f13-988f-bd3e2591037c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5290666b-725c-7b64-53b6-c8b95c3f9a76/disk-0.vmdk is in state: ready. {{(pid=63202) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 934.492661] env[63202]: ERROR oslo_vmware.rw_handles [None req-5423f696-c38b-4f13-988f-bd3e2591037c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5290666b-725c-7b64-53b6-c8b95c3f9a76/disk-0.vmdk due to incomplete transfer. [ 934.492891] env[63202]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-6f704a01-6976-4c12-b1e5-19d37c8750bd {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.501376] env[63202]: DEBUG oslo_vmware.rw_handles [None req-5423f696-c38b-4f13-988f-bd3e2591037c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5290666b-725c-7b64-53b6-c8b95c3f9a76/disk-0.vmdk. {{(pid=63202) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 934.501571] env[63202]: DEBUG nova.virt.vmwareapi.images [None req-5423f696-c38b-4f13-988f-bd3e2591037c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Uploaded image 376aee6c-1f98-4a60-858b-40c55f65b90e to the Glance image server {{(pid=63202) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 934.503907] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-5423f696-c38b-4f13-988f-bd3e2591037c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Destroying the VM {{(pid=63202) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 934.504165] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-9731ce0e-1f33-4afc-a587-189791a31281 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.509673] env[63202]: DEBUG oslo_vmware.api [None req-5423f696-c38b-4f13-988f-bd3e2591037c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for the task: (returnval){ [ 934.509673] env[63202]: value = "task-1385529" [ 934.509673] env[63202]: _type = "Task" [ 934.509673] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.517239] env[63202]: DEBUG oslo_vmware.api [None req-5423f696-c38b-4f13-988f-bd3e2591037c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385529, 'name': Destroy_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.558776] env[63202]: DEBUG oslo_vmware.api [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385523, 'name': PowerOnVM_Task, 'duration_secs': 1.051149} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.559057] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 934.559275] env[63202]: INFO nova.compute.manager [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Took 8.67 seconds to spawn the instance on the hypervisor. [ 934.559456] env[63202]: DEBUG nova.compute.manager [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 934.560505] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58407df3-bfde-49de-86c1-c1c5a92d380b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.608636] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385526, 'name': CreateVM_Task} progress is 25%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.629142] env[63202]: DEBUG nova.compute.manager [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 934.683588] env[63202]: DEBUG nova.network.neutron [req-fbee6873-ecff-48c4-a2da-f6079c4d6446 req-3c6a4385-3a68-4304-85c2-c32a8e79f696 service nova] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 934.802776] env[63202]: DEBUG nova.network.neutron [req-fbee6873-ecff-48c4-a2da-f6079c4d6446 req-3c6a4385-3a68-4304-85c2-c32a8e79f696 service nova] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.904023] env[63202]: DEBUG oslo_vmware.api [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': task-1385527, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.905296] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4bf0235-9720-45f1-975f-08773f8a691a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.915931] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-140746f4-084e-473a-bfd2-ca0aa5e50b3c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.945947] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5905ecaf-2ba2-40e9-b2a7-c0298bb18203 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.956471] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d0fc8d4-f5dd-4f17-8669-35b368987b62 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.971111] env[63202]: DEBUG nova.compute.provider_tree [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Updating inventory in ProviderTree for provider 79b33d17-3e75-494c-a550-67b275de2079 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 935.024586] env[63202]: DEBUG oslo_vmware.api [None req-5423f696-c38b-4f13-988f-bd3e2591037c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385529, 'name': Destroy_Task} progress is 100%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.083531] env[63202]: INFO nova.compute.manager [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Took 36.61 seconds to build instance. [ 935.112963] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385526, 'name': CreateVM_Task} progress is 25%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.306022] env[63202]: DEBUG oslo_concurrency.lockutils [req-fbee6873-ecff-48c4-a2da-f6079c4d6446 req-3c6a4385-3a68-4304-85c2-c32a8e79f696 service nova] Releasing lock "refresh_cache-8755bceb-d510-4429-bd98-d6a63faf739d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.306402] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquired lock "refresh_cache-8755bceb-d510-4429-bd98-d6a63faf739d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.306775] env[63202]: DEBUG nova.network.neutron [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 935.404435] env[63202]: DEBUG oslo_vmware.api [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': task-1385527, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.452688] env[63202]: DEBUG nova.objects.instance [None req-61daa1db-4e4a-4b19-a510-d2d715a52389 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lazy-loading 'flavor' on Instance uuid 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 935.495143] env[63202]: ERROR nova.scheduler.client.report [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [req-08de53c8-2cd1-4b6c-a348-68edfeaa4657] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 79b33d17-3e75-494c-a550-67b275de2079. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-08de53c8-2cd1-4b6c-a348-68edfeaa4657"}]} [ 935.516598] env[63202]: DEBUG nova.scheduler.client.report [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Refreshing inventories for resource provider 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 935.525515] env[63202]: DEBUG oslo_vmware.api [None req-5423f696-c38b-4f13-988f-bd3e2591037c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385529, 'name': Destroy_Task, 'duration_secs': 0.991383} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.525737] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-5423f696-c38b-4f13-988f-bd3e2591037c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Destroyed the VM [ 935.526279] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-5423f696-c38b-4f13-988f-bd3e2591037c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Deleting Snapshot of the VM instance {{(pid=63202) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 935.526599] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-ad1fff1b-c622-4ee1-9f41-2aec964383ce {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.532713] env[63202]: DEBUG oslo_vmware.api [None req-5423f696-c38b-4f13-988f-bd3e2591037c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for the task: (returnval){ [ 935.532713] env[63202]: value = "task-1385530" [ 935.532713] env[63202]: _type = "Task" [ 935.532713] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.533645] env[63202]: DEBUG nova.scheduler.client.report [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Updating ProviderTree inventory for provider 79b33d17-3e75-494c-a550-67b275de2079 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 935.533875] env[63202]: DEBUG nova.compute.provider_tree [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Updating inventory in ProviderTree for provider 79b33d17-3e75-494c-a550-67b275de2079 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 935.546977] env[63202]: DEBUG oslo_vmware.api [None req-5423f696-c38b-4f13-988f-bd3e2591037c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385530, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.548008] env[63202]: DEBUG nova.scheduler.client.report [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Refreshing aggregate associations for resource provider 79b33d17-3e75-494c-a550-67b275de2079, aggregates: None {{(pid=63202) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 935.565511] env[63202]: DEBUG nova.scheduler.client.report [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Refreshing trait associations for resource provider 79b33d17-3e75-494c-a550-67b275de2079, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,HW_ARCH_X86_64,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO {{(pid=63202) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 935.586227] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b010715-a501-4a2a-933b-3923c667c17d tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Lock "f3652744-e072-4700-80d4-b9eca414c5cb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 83.170s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.612646] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385526, 'name': CreateVM_Task, 'duration_secs': 2.471188} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.612822] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 935.613520] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.613693] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.614024] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 935.614290] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a2796e5b-069c-482f-8469-fbe858616137 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.621906] env[63202]: DEBUG oslo_vmware.api [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Waiting for the task: (returnval){ [ 935.621906] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]525d9edb-908b-adfa-c7e3-632b9c8ab0f4" [ 935.621906] env[63202]: _type = "Task" [ 935.621906] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.632489] env[63202]: DEBUG oslo_vmware.api [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]525d9edb-908b-adfa-c7e3-632b9c8ab0f4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.640824] env[63202]: DEBUG nova.compute.manager [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 935.667266] env[63202]: DEBUG nova.virt.hardware [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 935.667561] env[63202]: DEBUG nova.virt.hardware [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 935.667731] env[63202]: DEBUG nova.virt.hardware [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 935.667956] env[63202]: DEBUG nova.virt.hardware [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 935.668811] env[63202]: DEBUG nova.virt.hardware [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 935.668811] env[63202]: DEBUG nova.virt.hardware [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 935.668811] env[63202]: DEBUG nova.virt.hardware [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 935.668811] env[63202]: DEBUG nova.virt.hardware [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 935.669359] env[63202]: DEBUG nova.virt.hardware [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 935.669359] env[63202]: DEBUG nova.virt.hardware [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 935.669359] env[63202]: DEBUG nova.virt.hardware [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 935.670356] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-260e8308-7d29-49be-ab1f-3353ec378511 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.681048] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-343e6fc7-eeba-467b-bcb8-c71a166f6857 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.827073] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9b0c079-8c55-4142-9d48-1182f70b1430 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.835571] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f050230f-7410-4ab7-afd4-4c0bdfe03284 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.842285] env[63202]: DEBUG nova.network.neutron [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 935.870351] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-045a28d7-46fa-42ae-81db-cbab53de7c97 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.879021] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c6070d2-1188-4b87-955a-628a020cdf24 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.893440] env[63202]: DEBUG nova.compute.provider_tree [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Updating inventory in ProviderTree for provider 79b33d17-3e75-494c-a550-67b275de2079 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 935.909008] env[63202]: DEBUG oslo_vmware.api [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': task-1385527, 'name': PowerOnVM_Task, 'duration_secs': 2.124445} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.910502] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 935.910769] env[63202]: INFO nova.compute.manager [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Took 7.53 seconds to spawn the instance on the hypervisor. [ 935.910990] env[63202]: DEBUG nova.compute.manager [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 935.915217] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43d3323b-66f8-40f9-9403-ada7fbc62c2b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.918656] env[63202]: DEBUG oslo_concurrency.lockutils [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquiring lock "da285417-bd38-4387-8521-df326ca0e326" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.918881] env[63202]: DEBUG oslo_concurrency.lockutils [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lock "da285417-bd38-4387-8521-df326ca0e326" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.953288] env[63202]: DEBUG nova.network.neutron [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Successfully updated port: be85354c-2bf7-4c06-898f-7061cdfada23 {{(pid=63202) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 935.960825] env[63202]: DEBUG oslo_concurrency.lockutils [None req-61daa1db-4e4a-4b19-a510-d2d715a52389 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lock "0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.812s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.980728] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4eda49d7-5b02-4d51-99e9-f29b250eca15 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Acquiring lock "af2bedc2-28ee-4679-ae38-1cceb2af05d6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.981116] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4eda49d7-5b02-4d51-99e9-f29b250eca15 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Lock "af2bedc2-28ee-4679-ae38-1cceb2af05d6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.981243] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4eda49d7-5b02-4d51-99e9-f29b250eca15 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Acquiring lock "af2bedc2-28ee-4679-ae38-1cceb2af05d6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.981718] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4eda49d7-5b02-4d51-99e9-f29b250eca15 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Lock "af2bedc2-28ee-4679-ae38-1cceb2af05d6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.981718] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4eda49d7-5b02-4d51-99e9-f29b250eca15 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Lock "af2bedc2-28ee-4679-ae38-1cceb2af05d6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.985575] env[63202]: INFO nova.compute.manager [None req-4eda49d7-5b02-4d51-99e9-f29b250eca15 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Terminating instance [ 935.988759] env[63202]: DEBUG nova.compute.manager [None req-4eda49d7-5b02-4d51-99e9-f29b250eca15 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 935.988997] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-4eda49d7-5b02-4d51-99e9-f29b250eca15 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 935.991424] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24e097aa-1cd0-40dd-9630-9ca9921bae14 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.001823] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-4eda49d7-5b02-4d51-99e9-f29b250eca15 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 936.002515] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bd0a1731-5942-4aa0-8579-9b17c2bd6890 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.010015] env[63202]: DEBUG oslo_vmware.api [None req-4eda49d7-5b02-4d51-99e9-f29b250eca15 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Waiting for the task: (returnval){ [ 936.010015] env[63202]: value = "task-1385531" [ 936.010015] env[63202]: _type = "Task" [ 936.010015] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.019127] env[63202]: DEBUG oslo_vmware.api [None req-4eda49d7-5b02-4d51-99e9-f29b250eca15 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385531, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.045408] env[63202]: DEBUG oslo_vmware.api [None req-5423f696-c38b-4f13-988f-bd3e2591037c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385530, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.084419] env[63202]: DEBUG nova.network.neutron [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Updating instance_info_cache with network_info: [{"id": "faa30f43-2e38-4dbb-97ca-1b4c17565310", "address": "fa:16:3e:4e:1f:32", "network": {"id": "5a3b4c9b-2ca7-4f8b-8bbb-ea000db91402", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-543677750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f508ec04edc844a19640a8a85f27e5b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfaa30f43-2e", "ovs_interfaceid": "faa30f43-2e38-4dbb-97ca-1b4c17565310", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.133184] env[63202]: DEBUG oslo_vmware.api [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]525d9edb-908b-adfa-c7e3-632b9c8ab0f4, 'name': SearchDatastore_Task, 'duration_secs': 0.012202} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.133339] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.133593] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 936.133830] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.133977] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.134171] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 936.134781] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0b6435e5-261a-4a48-84dc-e7d2264566d8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.144039] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 936.144039] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 936.144791] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-07843fd2-e863-4c3b-aabf-ba12e91246b4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.150227] env[63202]: DEBUG oslo_vmware.api [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Waiting for the task: (returnval){ [ 936.150227] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52b51426-c8fa-4dce-b0cc-9b6763dac889" [ 936.150227] env[63202]: _type = "Task" [ 936.150227] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.161266] env[63202]: DEBUG oslo_vmware.api [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52b51426-c8fa-4dce-b0cc-9b6763dac889, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.164785] env[63202]: DEBUG nova.compute.manager [req-d356a9a6-73c7-4fe2-827a-6deb60f9a3c5 req-84b27cd6-19a1-4d9e-ae98-da2afcb08093 service nova] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Received event network-vif-plugged-be85354c-2bf7-4c06-898f-7061cdfada23 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 936.165172] env[63202]: DEBUG oslo_concurrency.lockutils [req-d356a9a6-73c7-4fe2-827a-6deb60f9a3c5 req-84b27cd6-19a1-4d9e-ae98-da2afcb08093 service nova] Acquiring lock "f3128c09-3680-4b0b-b463-3d6cd203fcf4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.165532] env[63202]: DEBUG oslo_concurrency.lockutils [req-d356a9a6-73c7-4fe2-827a-6deb60f9a3c5 req-84b27cd6-19a1-4d9e-ae98-da2afcb08093 service nova] Lock "f3128c09-3680-4b0b-b463-3d6cd203fcf4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.165831] env[63202]: DEBUG oslo_concurrency.lockutils [req-d356a9a6-73c7-4fe2-827a-6deb60f9a3c5 req-84b27cd6-19a1-4d9e-ae98-da2afcb08093 service nova] Lock "f3128c09-3680-4b0b-b463-3d6cd203fcf4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.166146] env[63202]: DEBUG nova.compute.manager [req-d356a9a6-73c7-4fe2-827a-6deb60f9a3c5 req-84b27cd6-19a1-4d9e-ae98-da2afcb08093 service nova] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] No waiting events found dispatching network-vif-plugged-be85354c-2bf7-4c06-898f-7061cdfada23 {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 936.166450] env[63202]: WARNING nova.compute.manager [req-d356a9a6-73c7-4fe2-827a-6deb60f9a3c5 req-84b27cd6-19a1-4d9e-ae98-da2afcb08093 service nova] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Received unexpected event network-vif-plugged-be85354c-2bf7-4c06-898f-7061cdfada23 for instance with vm_state building and task_state spawning. [ 936.166747] env[63202]: DEBUG nova.compute.manager [req-d356a9a6-73c7-4fe2-827a-6deb60f9a3c5 req-84b27cd6-19a1-4d9e-ae98-da2afcb08093 service nova] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Received event network-changed-be85354c-2bf7-4c06-898f-7061cdfada23 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 936.167078] env[63202]: DEBUG nova.compute.manager [req-d356a9a6-73c7-4fe2-827a-6deb60f9a3c5 req-84b27cd6-19a1-4d9e-ae98-da2afcb08093 service nova] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Refreshing instance network info cache due to event network-changed-be85354c-2bf7-4c06-898f-7061cdfada23. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 936.167408] env[63202]: DEBUG oslo_concurrency.lockutils [req-d356a9a6-73c7-4fe2-827a-6deb60f9a3c5 req-84b27cd6-19a1-4d9e-ae98-da2afcb08093 service nova] Acquiring lock "refresh_cache-f3128c09-3680-4b0b-b463-3d6cd203fcf4" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.167672] env[63202]: DEBUG oslo_concurrency.lockutils [req-d356a9a6-73c7-4fe2-827a-6deb60f9a3c5 req-84b27cd6-19a1-4d9e-ae98-da2afcb08093 service nova] Acquired lock "refresh_cache-f3128c09-3680-4b0b-b463-3d6cd203fcf4" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.167976] env[63202]: DEBUG nova.network.neutron [req-d356a9a6-73c7-4fe2-827a-6deb60f9a3c5 req-84b27cd6-19a1-4d9e-ae98-da2afcb08093 service nova] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Refreshing network info cache for port be85354c-2bf7-4c06-898f-7061cdfada23 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 936.421530] env[63202]: DEBUG nova.compute.manager [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: da285417-bd38-4387-8521-df326ca0e326] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 936.434762] env[63202]: DEBUG nova.scheduler.client.report [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Updated inventory for provider 79b33d17-3e75-494c-a550-67b275de2079 with generation 93 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 936.434762] env[63202]: DEBUG nova.compute.provider_tree [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Updating resource provider 79b33d17-3e75-494c-a550-67b275de2079 generation from 93 to 94 during operation: update_inventory {{(pid=63202) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 936.434762] env[63202]: DEBUG nova.compute.provider_tree [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Updating inventory in ProviderTree for provider 79b33d17-3e75-494c-a550-67b275de2079 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 936.437594] env[63202]: INFO nova.compute.manager [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Took 37.13 seconds to build instance. [ 936.460041] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Acquiring lock "refresh_cache-f3128c09-3680-4b0b-b463-3d6cd203fcf4" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.520992] env[63202]: DEBUG oslo_vmware.api [None req-4eda49d7-5b02-4d51-99e9-f29b250eca15 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385531, 'name': PowerOffVM_Task, 'duration_secs': 0.322506} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.521325] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-4eda49d7-5b02-4d51-99e9-f29b250eca15 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 936.522028] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-4eda49d7-5b02-4d51-99e9-f29b250eca15 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 936.522028] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d018535b-994b-4401-a193-1985517c681c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.546866] env[63202]: DEBUG oslo_vmware.api [None req-5423f696-c38b-4f13-988f-bd3e2591037c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385530, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.588229] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-4eda49d7-5b02-4d51-99e9-f29b250eca15 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 936.588697] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-4eda49d7-5b02-4d51-99e9-f29b250eca15 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Deleting contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 936.588697] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-4eda49d7-5b02-4d51-99e9-f29b250eca15 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Deleting the datastore file [datastore1] af2bedc2-28ee-4679-ae38-1cceb2af05d6 {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 936.589262] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Releasing lock "refresh_cache-8755bceb-d510-4429-bd98-d6a63faf739d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.589674] env[63202]: DEBUG nova.compute.manager [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Instance network_info: |[{"id": "faa30f43-2e38-4dbb-97ca-1b4c17565310", "address": "fa:16:3e:4e:1f:32", "network": {"id": "5a3b4c9b-2ca7-4f8b-8bbb-ea000db91402", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-543677750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f508ec04edc844a19640a8a85f27e5b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfaa30f43-2e", "ovs_interfaceid": "faa30f43-2e38-4dbb-97ca-1b4c17565310", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 936.589990] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3e73c87e-798c-4cd8-9bf2-f9755743023f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.593208] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4e:1f:32', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'faa30f43-2e38-4dbb-97ca-1b4c17565310', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 936.606223] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Creating folder: Project (f508ec04edc844a19640a8a85f27e5b9). Parent ref: group-v294090. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 936.606812] env[63202]: DEBUG oslo_concurrency.lockutils [None req-40459ef6-272f-4712-84fb-409ed0898d79 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Acquiring lock "0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.607368] env[63202]: DEBUG oslo_concurrency.lockutils [None req-40459ef6-272f-4712-84fb-409ed0898d79 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lock "0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.607512] env[63202]: DEBUG nova.compute.manager [None req-40459ef6-272f-4712-84fb-409ed0898d79 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 936.608255] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-07a7ed05-d7d9-4161-9180-d3a19b67aaab {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.611748] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de754d25-20ae-483e-b96a-4c099f062bdf {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.626430] env[63202]: DEBUG nova.compute.manager [None req-40459ef6-272f-4712-84fb-409ed0898d79 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63202) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 936.627073] env[63202]: DEBUG nova.objects.instance [None req-40459ef6-272f-4712-84fb-409ed0898d79 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lazy-loading 'flavor' on Instance uuid 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 936.629814] env[63202]: DEBUG oslo_vmware.api [None req-4eda49d7-5b02-4d51-99e9-f29b250eca15 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Waiting for the task: (returnval){ [ 936.629814] env[63202]: value = "task-1385533" [ 936.629814] env[63202]: _type = "Task" [ 936.629814] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.631373] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Created folder: Project (f508ec04edc844a19640a8a85f27e5b9) in parent group-v294090. [ 936.631631] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Creating folder: Instances. Parent ref: group-v294159. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 936.634973] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b752cf94-4ac6-4513-bc2e-8babbc26d055 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.644044] env[63202]: DEBUG oslo_vmware.api [None req-4eda49d7-5b02-4d51-99e9-f29b250eca15 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385533, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.646110] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Created folder: Instances in parent group-v294159. [ 936.646360] env[63202]: DEBUG oslo.service.loopingcall [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 936.646559] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 936.646848] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-da94de27-f6a4-4744-a129-909cd2c0beb2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.673725] env[63202]: DEBUG oslo_vmware.api [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52b51426-c8fa-4dce-b0cc-9b6763dac889, 'name': SearchDatastore_Task, 'duration_secs': 0.011562} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.675656] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 936.675656] env[63202]: value = "task-1385536" [ 936.675656] env[63202]: _type = "Task" [ 936.675656] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.676206] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e5688527-c4bc-451f-bcfb-0eafc6c871fa {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.684224] env[63202]: DEBUG oslo_vmware.api [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Waiting for the task: (returnval){ [ 936.684224] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52adb643-b516-5254-9f6b-11b00179150d" [ 936.684224] env[63202]: _type = "Task" [ 936.684224] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.687639] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385536, 'name': CreateVM_Task} progress is 6%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.695085] env[63202]: DEBUG oslo_vmware.api [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52adb643-b516-5254-9f6b-11b00179150d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.721658] env[63202]: DEBUG nova.network.neutron [req-d356a9a6-73c7-4fe2-827a-6deb60f9a3c5 req-84b27cd6-19a1-4d9e-ae98-da2afcb08093 service nova] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 936.862031] env[63202]: DEBUG nova.network.neutron [req-d356a9a6-73c7-4fe2-827a-6deb60f9a3c5 req-84b27cd6-19a1-4d9e-ae98-da2afcb08093 service nova] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.938722] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.325s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.939275] env[63202]: DEBUG nova.compute.manager [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 936.942608] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 15.473s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.943994] env[63202]: DEBUG oslo_concurrency.lockutils [None req-64c72a94-2101-40bf-880d-0c5177f86aeb tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Lock "fc7d3924-a624-4ea4-890c-7628595fb733" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 78.998s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.951674] env[63202]: DEBUG oslo_concurrency.lockutils [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.051727] env[63202]: DEBUG oslo_vmware.api [None req-5423f696-c38b-4f13-988f-bd3e2591037c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385530, 'name': RemoveSnapshot_Task, 'duration_secs': 1.322567} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.051727] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-5423f696-c38b-4f13-988f-bd3e2591037c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Deleted Snapshot of the VM instance {{(pid=63202) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 937.051956] env[63202]: INFO nova.compute.manager [None req-5423f696-c38b-4f13-988f-bd3e2591037c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Took 18.29 seconds to snapshot the instance on the hypervisor. [ 937.138552] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-40459ef6-272f-4712-84fb-409ed0898d79 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 937.138923] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d31b5b0d-e049-48d0-ae37-a205f7526d33 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.145820] env[63202]: DEBUG oslo_vmware.api [None req-4eda49d7-5b02-4d51-99e9-f29b250eca15 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385533, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.219283} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.147050] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-4eda49d7-5b02-4d51-99e9-f29b250eca15 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 937.147245] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-4eda49d7-5b02-4d51-99e9-f29b250eca15 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Deleted contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 937.147425] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-4eda49d7-5b02-4d51-99e9-f29b250eca15 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 937.147615] env[63202]: INFO nova.compute.manager [None req-4eda49d7-5b02-4d51-99e9-f29b250eca15 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Took 1.16 seconds to destroy the instance on the hypervisor. [ 937.147877] env[63202]: DEBUG oslo.service.loopingcall [None req-4eda49d7-5b02-4d51-99e9-f29b250eca15 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 937.148161] env[63202]: DEBUG oslo_vmware.api [None req-40459ef6-272f-4712-84fb-409ed0898d79 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Waiting for the task: (returnval){ [ 937.148161] env[63202]: value = "task-1385537" [ 937.148161] env[63202]: _type = "Task" [ 937.148161] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.148408] env[63202]: DEBUG nova.compute.manager [-] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 937.148470] env[63202]: DEBUG nova.network.neutron [-] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 937.158518] env[63202]: DEBUG oslo_vmware.api [None req-40459ef6-272f-4712-84fb-409ed0898d79 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385537, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.190986] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385536, 'name': CreateVM_Task} progress is 99%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.201438] env[63202]: DEBUG oslo_vmware.api [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52adb643-b516-5254-9f6b-11b00179150d, 'name': SearchDatastore_Task, 'duration_secs': 0.017503} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.201754] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.202037] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore2] d0d6e380-9337-4f69-8434-6a33ac8a33d6/d0d6e380-9337-4f69-8434-6a33ac8a33d6.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 937.202350] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4a732ced-5966-4744-865b-19eda559f47f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.209961] env[63202]: DEBUG oslo_vmware.api [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Waiting for the task: (returnval){ [ 937.209961] env[63202]: value = "task-1385538" [ 937.209961] env[63202]: _type = "Task" [ 937.209961] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.221123] env[63202]: DEBUG oslo_vmware.api [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Task: {'id': task-1385538, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.365133] env[63202]: DEBUG oslo_concurrency.lockutils [req-d356a9a6-73c7-4fe2-827a-6deb60f9a3c5 req-84b27cd6-19a1-4d9e-ae98-da2afcb08093 service nova] Releasing lock "refresh_cache-f3128c09-3680-4b0b-b463-3d6cd203fcf4" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.365646] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Acquired lock "refresh_cache-f3128c09-3680-4b0b-b463-3d6cd203fcf4" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.365789] env[63202]: DEBUG nova.network.neutron [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 937.446438] env[63202]: DEBUG nova.compute.utils [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 937.452226] env[63202]: INFO nova.compute.claims [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 937.458443] env[63202]: DEBUG nova.compute.manager [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 937.458443] env[63202]: DEBUG nova.network.neutron [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 937.520885] env[63202]: DEBUG nova.policy [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '63a29c785cf240a7b6418dadc7119574', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f7d28fea097541adadf0839940568409', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 937.661404] env[63202]: DEBUG oslo_vmware.api [None req-40459ef6-272f-4712-84fb-409ed0898d79 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385537, 'name': PowerOffVM_Task, 'duration_secs': 0.215403} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.661762] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-40459ef6-272f-4712-84fb-409ed0898d79 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 937.661881] env[63202]: DEBUG nova.compute.manager [None req-40459ef6-272f-4712-84fb-409ed0898d79 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 937.663051] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f454abd7-5310-4c76-bc45-3a61e424638a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.689919] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385536, 'name': CreateVM_Task, 'duration_secs': 0.577978} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.690708] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 937.691494] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.691793] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.692185] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 937.692500] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-45c27e8a-37a4-4c13-9963-b7c849f2fe1a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.698305] env[63202]: DEBUG oslo_vmware.api [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 937.698305] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52128ede-b480-a287-6978-7163ab19aef2" [ 937.698305] env[63202]: _type = "Task" [ 937.698305] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.707576] env[63202]: DEBUG oslo_vmware.api [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52128ede-b480-a287-6978-7163ab19aef2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.720720] env[63202]: DEBUG oslo_vmware.api [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Task: {'id': task-1385538, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.762183] env[63202]: DEBUG oslo_concurrency.lockutils [None req-02dae3ed-6d57-4066-a061-72500d73ac65 tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Acquiring lock "2243bb69-0dc5-49cd-b94e-73e703cbadc0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.762183] env[63202]: DEBUG oslo_concurrency.lockutils [None req-02dae3ed-6d57-4066-a061-72500d73ac65 tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Lock "2243bb69-0dc5-49cd-b94e-73e703cbadc0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.762183] env[63202]: DEBUG oslo_concurrency.lockutils [None req-02dae3ed-6d57-4066-a061-72500d73ac65 tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Acquiring lock "2243bb69-0dc5-49cd-b94e-73e703cbadc0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.762183] env[63202]: DEBUG oslo_concurrency.lockutils [None req-02dae3ed-6d57-4066-a061-72500d73ac65 tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Lock "2243bb69-0dc5-49cd-b94e-73e703cbadc0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.762447] env[63202]: DEBUG oslo_concurrency.lockutils [None req-02dae3ed-6d57-4066-a061-72500d73ac65 tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Lock "2243bb69-0dc5-49cd-b94e-73e703cbadc0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.764459] env[63202]: INFO nova.compute.manager [None req-02dae3ed-6d57-4066-a061-72500d73ac65 tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Terminating instance [ 937.766576] env[63202]: DEBUG nova.compute.manager [None req-02dae3ed-6d57-4066-a061-72500d73ac65 tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 937.766787] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-02dae3ed-6d57-4066-a061-72500d73ac65 tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 937.767654] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a91bda0b-1efb-4a00-88d0-46c8bba712ec {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.775804] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-02dae3ed-6d57-4066-a061-72500d73ac65 tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 937.776280] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-19ba85cb-8d10-4cb5-bdb1-2eabd9dfecc5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.783309] env[63202]: DEBUG oslo_vmware.api [None req-02dae3ed-6d57-4066-a061-72500d73ac65 tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Waiting for the task: (returnval){ [ 937.783309] env[63202]: value = "task-1385539" [ 937.783309] env[63202]: _type = "Task" [ 937.783309] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.792152] env[63202]: DEBUG oslo_vmware.api [None req-02dae3ed-6d57-4066-a061-72500d73ac65 tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Task: {'id': task-1385539, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.809160] env[63202]: INFO nova.compute.manager [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Rebuilding instance [ 937.839243] env[63202]: DEBUG nova.compute.manager [req-fb0b241e-e744-43ed-8d52-4672530db607 req-57ca265b-72e0-4aee-bc63-7428917e4310 service nova] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Received event network-vif-deleted-f4c43c2b-ec12-47f7-84d8-4190afbf1b7d {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 937.839243] env[63202]: INFO nova.compute.manager [req-fb0b241e-e744-43ed-8d52-4672530db607 req-57ca265b-72e0-4aee-bc63-7428917e4310 service nova] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Neutron deleted interface f4c43c2b-ec12-47f7-84d8-4190afbf1b7d; detaching it from the instance and deleting it from the info cache [ 937.843417] env[63202]: DEBUG nova.network.neutron [req-fb0b241e-e744-43ed-8d52-4672530db607 req-57ca265b-72e0-4aee-bc63-7428917e4310 service nova] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.843417] env[63202]: DEBUG nova.network.neutron [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Successfully created port: 5eeb4ae8-9356-402a-965d-224788f04129 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 937.879028] env[63202]: DEBUG nova.compute.manager [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 937.882070] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94b819f4-bc76-45fd-afcf-0bb3f2bdf67e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.926896] env[63202]: DEBUG nova.network.neutron [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 937.956378] env[63202]: DEBUG nova.compute.manager [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 937.961629] env[63202]: INFO nova.compute.resource_tracker [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Updating resource usage from migration 78063a2a-a85c-4117-89c4-77c0fd76a7ae [ 937.978277] env[63202]: DEBUG nova.network.neutron [-] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.177213] env[63202]: DEBUG oslo_concurrency.lockutils [None req-40459ef6-272f-4712-84fb-409ed0898d79 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lock "0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.570s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.212105] env[63202]: DEBUG oslo_vmware.api [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52128ede-b480-a287-6978-7163ab19aef2, 'name': SearchDatastore_Task, 'duration_secs': 0.065309} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.217403] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.217403] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 938.217403] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.217403] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.217679] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 938.219622] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-088322d5-f771-40be-beb6-535396d14593 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.228077] env[63202]: DEBUG oslo_vmware.api [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Task: {'id': task-1385538, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.797891} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.228077] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore2] d0d6e380-9337-4f69-8434-6a33ac8a33d6/d0d6e380-9337-4f69-8434-6a33ac8a33d6.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 938.228077] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 938.228077] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a1e35765-4ecd-471c-ba80-4759756aa8e1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.230788] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 938.230996] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 938.232745] env[63202]: DEBUG nova.network.neutron [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Updating instance_info_cache with network_info: [{"id": "be85354c-2bf7-4c06-898f-7061cdfada23", "address": "fa:16:3e:e6:8e:d3", "network": {"id": "8a3e2159-c8ca-4ad6-b9c6-87029c42fa61", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-503250460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "417f815aec0c451a8ccd713ad6c64ad2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbe85354c-2b", "ovs_interfaceid": "be85354c-2bf7-4c06-898f-7061cdfada23", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.236964] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-14fe3158-d32c-4002-bd26-200ff8856265 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.241491] env[63202]: DEBUG oslo_vmware.api [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Waiting for the task: (returnval){ [ 938.241491] env[63202]: value = "task-1385540" [ 938.241491] env[63202]: _type = "Task" [ 938.241491] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.247065] env[63202]: DEBUG oslo_vmware.api [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 938.247065] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52aab4e7-89db-36dc-2957-ed569db6d0ac" [ 938.247065] env[63202]: _type = "Task" [ 938.247065] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.256596] env[63202]: DEBUG oslo_vmware.api [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Task: {'id': task-1385540, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.262817] env[63202]: DEBUG oslo_vmware.api [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52aab4e7-89db-36dc-2957-ed569db6d0ac, 'name': SearchDatastore_Task, 'duration_secs': 0.010828} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.266345] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4014d54-116f-4414-b3b3-869d9511cf82 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.271096] env[63202]: DEBUG oslo_vmware.api [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 938.271096] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]5286d88d-c450-4d94-f45f-a2527a28eabc" [ 938.271096] env[63202]: _type = "Task" [ 938.271096] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.279322] env[63202]: DEBUG oslo_vmware.api [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]5286d88d-c450-4d94-f45f-a2527a28eabc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.292135] env[63202]: DEBUG oslo_vmware.api [None req-02dae3ed-6d57-4066-a061-72500d73ac65 tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Task: {'id': task-1385539, 'name': PowerOffVM_Task, 'duration_secs': 0.37562} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.292426] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-02dae3ed-6d57-4066-a061-72500d73ac65 tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 938.292599] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-02dae3ed-6d57-4066-a061-72500d73ac65 tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 938.293254] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-45fde5cb-7583-4c82-828e-56ca91b772b6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.343578] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad06d305-ea41-4d92-af9d-484295a699af {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.346498] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d91d3089-5efa-4ee5-b58e-90420aa7fd07 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.353746] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e24a2bc4-d8e6-440a-ba27-89ca74353817 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.362350] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f50b46e8-1249-425a-afb7-1f10818b95e2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.403947] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 938.413386] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bed129d7-e704-4226-aa09-b77a20e6a1f0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.415431] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8beb0412-86de-4f82-b049-491f20f11156 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.417763] env[63202]: DEBUG nova.compute.manager [req-fb0b241e-e744-43ed-8d52-4672530db607 req-57ca265b-72e0-4aee-bc63-7428917e4310 service nova] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Detach interface failed, port_id=f4c43c2b-ec12-47f7-84d8-4190afbf1b7d, reason: Instance af2bedc2-28ee-4679-ae38-1cceb2af05d6 could not be found. {{(pid=63202) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 938.423955] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f01066d0-7829-49cd-9039-0850cb06256e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.427446] env[63202]: DEBUG oslo_vmware.api [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Waiting for the task: (returnval){ [ 938.427446] env[63202]: value = "task-1385542" [ 938.427446] env[63202]: _type = "Task" [ 938.427446] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.437811] env[63202]: DEBUG nova.compute.provider_tree [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 938.444630] env[63202]: DEBUG oslo_vmware.api [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': task-1385542, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.481220] env[63202]: INFO nova.compute.manager [-] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Took 1.33 seconds to deallocate network for instance. [ 938.742685] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Releasing lock "refresh_cache-f3128c09-3680-4b0b-b463-3d6cd203fcf4" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.743028] env[63202]: DEBUG nova.compute.manager [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Instance network_info: |[{"id": "be85354c-2bf7-4c06-898f-7061cdfada23", "address": "fa:16:3e:e6:8e:d3", "network": {"id": "8a3e2159-c8ca-4ad6-b9c6-87029c42fa61", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-503250460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "417f815aec0c451a8ccd713ad6c64ad2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbe85354c-2b", "ovs_interfaceid": "be85354c-2bf7-4c06-898f-7061cdfada23", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 938.747018] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e6:8e:d3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1002b79b-224e-41e3-a484-4245a767147a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'be85354c-2bf7-4c06-898f-7061cdfada23', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 938.751030] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Creating folder: Project (417f815aec0c451a8ccd713ad6c64ad2). Parent ref: group-v294090. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 938.756529] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-757b3ac1-47a2-40cf-8359-898a60d8fe0d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.764263] env[63202]: DEBUG oslo_vmware.api [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Task: {'id': task-1385540, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062229} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.764447] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 938.766173] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8d1f170-877a-4601-a045-cbd44587be4d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.769226] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Created folder: Project (417f815aec0c451a8ccd713ad6c64ad2) in parent group-v294090. [ 938.769226] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Creating folder: Instances. Parent ref: group-v294162. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 938.769226] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3049c57d-5722-4a3c-b2a4-afcddcf65754 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.794837] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] d0d6e380-9337-4f69-8434-6a33ac8a33d6/d0d6e380-9337-4f69-8434-6a33ac8a33d6.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 938.798546] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2b1e069d-62fb-4687-a5ed-bedb74438a00 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.812056] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Created folder: Instances in parent group-v294162. [ 938.812351] env[63202]: DEBUG oslo.service.loopingcall [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 938.814155] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 938.814768] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c3a09aa3-b75e-47c5-b1bb-8eb19cf220bd {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.833359] env[63202]: DEBUG oslo_vmware.api [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Waiting for the task: (returnval){ [ 938.833359] env[63202]: value = "task-1385545" [ 938.833359] env[63202]: _type = "Task" [ 938.833359] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.837927] env[63202]: DEBUG oslo_vmware.api [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]5286d88d-c450-4d94-f45f-a2527a28eabc, 'name': SearchDatastore_Task, 'duration_secs': 0.010633} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.839306] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.839588] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore2] 8755bceb-d510-4429-bd98-d6a63faf739d/8755bceb-d510-4429-bd98-d6a63faf739d.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 938.839827] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 938.839827] env[63202]: value = "task-1385546" [ 938.839827] env[63202]: _type = "Task" [ 938.839827] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.840369] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a1a74871-be03-49ee-af5d-3194d1d602ec {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.848283] env[63202]: DEBUG oslo_vmware.api [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Task: {'id': task-1385545, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.854177] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385546, 'name': CreateVM_Task} progress is 6%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.855474] env[63202]: DEBUG oslo_vmware.api [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 938.855474] env[63202]: value = "task-1385547" [ 938.855474] env[63202]: _type = "Task" [ 938.855474] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.864848] env[63202]: DEBUG oslo_vmware.api [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385547, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.936865] env[63202]: DEBUG oslo_vmware.api [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': task-1385542, 'name': PowerOffVM_Task, 'duration_secs': 0.38964} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.937162] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 938.937385] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 938.938166] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79b9207d-66ed-47d8-8fe5-cbf95e8a3466 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.941338] env[63202]: DEBUG nova.scheduler.client.report [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 938.948507] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 938.948749] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5be98217-ebb4-4ebb-b198-32a9f558c8ff {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.971272] env[63202]: DEBUG nova.compute.manager [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 938.975495] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 938.975716] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Deleting contents of the VM from datastore datastore2 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 938.975897] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Deleting the datastore file [datastore2] fc7d3924-a624-4ea4-890c-7628595fb733 {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 938.976179] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3c258841-7ef6-45ef-af4c-2ff0631ee6d4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.982407] env[63202]: DEBUG oslo_vmware.api [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Waiting for the task: (returnval){ [ 938.982407] env[63202]: value = "task-1385549" [ 938.982407] env[63202]: _type = "Task" [ 938.982407] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.987592] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4eda49d7-5b02-4d51-99e9-f29b250eca15 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.992987] env[63202]: DEBUG oslo_vmware.api [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': task-1385549, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.001689] env[63202]: DEBUG nova.virt.hardware [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 939.002215] env[63202]: DEBUG nova.virt.hardware [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 939.002215] env[63202]: DEBUG nova.virt.hardware [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 939.002326] env[63202]: DEBUG nova.virt.hardware [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 939.002436] env[63202]: DEBUG nova.virt.hardware [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 939.002657] env[63202]: DEBUG nova.virt.hardware [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 939.002877] env[63202]: DEBUG nova.virt.hardware [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 939.003060] env[63202]: DEBUG nova.virt.hardware [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 939.003235] env[63202]: DEBUG nova.virt.hardware [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 939.003403] env[63202]: DEBUG nova.virt.hardware [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 939.003574] env[63202]: DEBUG nova.virt.hardware [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 939.004422] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b87a42f-9971-4301-8d4c-524150d6e1b2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.011980] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8618778c-0189-4775-a99b-599152f4544f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.153450] env[63202]: DEBUG nova.objects.instance [None req-8ba75055-34c9-4e5c-a6f6-df51e7ac604b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lazy-loading 'flavor' on Instance uuid 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 939.253288] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 939.305820] env[63202]: DEBUG nova.compute.manager [req-23bc6a5b-3271-4a26-807f-d216f4d156ba req-d3eb82ad-3097-4ceb-bd0c-4c11345db14a service nova] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Received event network-vif-plugged-5eeb4ae8-9356-402a-965d-224788f04129 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 939.305820] env[63202]: DEBUG oslo_concurrency.lockutils [req-23bc6a5b-3271-4a26-807f-d216f4d156ba req-d3eb82ad-3097-4ceb-bd0c-4c11345db14a service nova] Acquiring lock "7437595c-fa35-483e-95f3-b75405b6bd13-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.306076] env[63202]: DEBUG oslo_concurrency.lockutils [req-23bc6a5b-3271-4a26-807f-d216f4d156ba req-d3eb82ad-3097-4ceb-bd0c-4c11345db14a service nova] Lock "7437595c-fa35-483e-95f3-b75405b6bd13-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.306199] env[63202]: DEBUG oslo_concurrency.lockutils [req-23bc6a5b-3271-4a26-807f-d216f4d156ba req-d3eb82ad-3097-4ceb-bd0c-4c11345db14a service nova] Lock "7437595c-fa35-483e-95f3-b75405b6bd13-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.306366] env[63202]: DEBUG nova.compute.manager [req-23bc6a5b-3271-4a26-807f-d216f4d156ba req-d3eb82ad-3097-4ceb-bd0c-4c11345db14a service nova] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] No waiting events found dispatching network-vif-plugged-5eeb4ae8-9356-402a-965d-224788f04129 {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 939.306584] env[63202]: WARNING nova.compute.manager [req-23bc6a5b-3271-4a26-807f-d216f4d156ba req-d3eb82ad-3097-4ceb-bd0c-4c11345db14a service nova] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Received unexpected event network-vif-plugged-5eeb4ae8-9356-402a-965d-224788f04129 for instance with vm_state building and task_state spawning. [ 939.342642] env[63202]: DEBUG oslo_vmware.api [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Task: {'id': task-1385545, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.350601] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385546, 'name': CreateVM_Task} progress is 25%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.368835] env[63202]: DEBUG oslo_vmware.api [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385547, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.411433] env[63202]: DEBUG nova.network.neutron [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Successfully updated port: 5eeb4ae8-9356-402a-965d-224788f04129 {{(pid=63202) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 939.446970] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.504s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.447273] env[63202]: INFO nova.compute.manager [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Migrating [ 939.447538] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.447719] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquired lock "compute-rpcapi-router" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.449383] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.622s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.451346] env[63202]: INFO nova.compute.claims [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 939.455240] env[63202]: INFO nova.compute.rpcapi [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 939.455784] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Releasing lock "compute-rpcapi-router" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.494172] env[63202]: DEBUG oslo_vmware.api [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': task-1385549, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.336691} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.494451] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 939.494715] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Deleted contents of the VM from datastore datastore2 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 939.494904] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 939.658445] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8ba75055-34c9-4e5c-a6f6-df51e7ac604b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Acquiring lock "refresh_cache-0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.658664] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8ba75055-34c9-4e5c-a6f6-df51e7ac604b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Acquired lock "refresh_cache-0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.658808] env[63202]: DEBUG nova.network.neutron [None req-8ba75055-34c9-4e5c-a6f6-df51e7ac604b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 939.658983] env[63202]: DEBUG nova.objects.instance [None req-8ba75055-34c9-4e5c-a6f6-df51e7ac604b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lazy-loading 'info_cache' on Instance uuid 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 939.842449] env[63202]: DEBUG oslo_vmware.api [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Task: {'id': task-1385545, 'name': ReconfigVM_Task, 'duration_secs': 0.937191} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.842743] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Reconfigured VM instance instance-0000004e to attach disk [datastore2] d0d6e380-9337-4f69-8434-6a33ac8a33d6/d0d6e380-9337-4f69-8434-6a33ac8a33d6.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 939.843562] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ccf8cba4-07c0-4673-9b13-5a247d929216 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.853165] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385546, 'name': CreateVM_Task} progress is 25%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.854390] env[63202]: DEBUG oslo_vmware.api [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Waiting for the task: (returnval){ [ 939.854390] env[63202]: value = "task-1385550" [ 939.854390] env[63202]: _type = "Task" [ 939.854390] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.864330] env[63202]: DEBUG oslo_vmware.api [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Task: {'id': task-1385550, 'name': Rename_Task} progress is 6%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.867215] env[63202]: DEBUG oslo_vmware.api [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385547, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.858899} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.867450] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore2] 8755bceb-d510-4429-bd98-d6a63faf739d/8755bceb-d510-4429-bd98-d6a63faf739d.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 939.867676] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 939.867886] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ba1874a4-3af3-4d79-97da-5cec4f5e107b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.873065] env[63202]: DEBUG oslo_vmware.api [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 939.873065] env[63202]: value = "task-1385551" [ 939.873065] env[63202]: _type = "Task" [ 939.873065] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.880295] env[63202]: DEBUG oslo_vmware.api [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385551, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.916220] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquiring lock "refresh_cache-7437595c-fa35-483e-95f3-b75405b6bd13" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.916574] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquired lock "refresh_cache-7437595c-fa35-483e-95f3-b75405b6bd13" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.916637] env[63202]: DEBUG nova.network.neutron [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 939.975981] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "refresh_cache-e775e5e1-521a-4fc7-80e6-bcb6a70516c5" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.975981] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquired lock "refresh_cache-e775e5e1-521a-4fc7-80e6-bcb6a70516c5" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.975981] env[63202]: DEBUG nova.network.neutron [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 940.162471] env[63202]: DEBUG nova.objects.base [None req-8ba75055-34c9-4e5c-a6f6-df51e7ac604b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Object Instance<0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb> lazy-loaded attributes: flavor,info_cache {{(pid=63202) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 940.254167] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 940.353632] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385546, 'name': CreateVM_Task} progress is 25%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.362304] env[63202]: DEBUG oslo_vmware.api [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Task: {'id': task-1385550, 'name': Rename_Task, 'duration_secs': 0.227436} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.362592] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 940.362833] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f9d008f2-f1ac-4627-8259-6839f9421c16 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.368886] env[63202]: DEBUG oslo_vmware.api [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Waiting for the task: (returnval){ [ 940.368886] env[63202]: value = "task-1385552" [ 940.368886] env[63202]: _type = "Task" [ 940.368886] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.377539] env[63202]: DEBUG oslo_vmware.api [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Task: {'id': task-1385552, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.385381] env[63202]: DEBUG oslo_vmware.api [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385551, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064196} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.385630] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 940.386467] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b1a8f6d-0830-404f-b6a7-06eff13796c3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.408371] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Reconfiguring VM instance instance-0000004f to attach disk [datastore2] 8755bceb-d510-4429-bd98-d6a63faf739d/8755bceb-d510-4429-bd98-d6a63faf739d.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 940.408657] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-052c303e-5cea-4294-9374-1df211aa16f7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.429694] env[63202]: DEBUG oslo_vmware.api [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 940.429694] env[63202]: value = "task-1385553" [ 940.429694] env[63202]: _type = "Task" [ 940.429694] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.437902] env[63202]: DEBUG oslo_vmware.api [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385553, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.452127] env[63202]: DEBUG nova.network.neutron [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 940.490422] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-02dae3ed-6d57-4066-a061-72500d73ac65 tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 940.490573] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-02dae3ed-6d57-4066-a061-72500d73ac65 tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Deleting contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 940.490765] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-02dae3ed-6d57-4066-a061-72500d73ac65 tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Deleting the datastore file [datastore1] 2243bb69-0dc5-49cd-b94e-73e703cbadc0 {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 940.491038] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-14c344ce-2a87-493a-a919-d710da1ef14a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.501933] env[63202]: DEBUG oslo_vmware.api [None req-02dae3ed-6d57-4066-a061-72500d73ac65 tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Waiting for the task: (returnval){ [ 940.501933] env[63202]: value = "task-1385554" [ 940.501933] env[63202]: _type = "Task" [ 940.501933] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.517058] env[63202]: DEBUG oslo_vmware.api [None req-02dae3ed-6d57-4066-a061-72500d73ac65 tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Task: {'id': task-1385554, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.538588] env[63202]: DEBUG nova.virt.hardware [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 940.538843] env[63202]: DEBUG nova.virt.hardware [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 940.539022] env[63202]: DEBUG nova.virt.hardware [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 940.539215] env[63202]: DEBUG nova.virt.hardware [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 940.539359] env[63202]: DEBUG nova.virt.hardware [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 940.539514] env[63202]: DEBUG nova.virt.hardware [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 940.539746] env[63202]: DEBUG nova.virt.hardware [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 940.539913] env[63202]: DEBUG nova.virt.hardware [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 940.540097] env[63202]: DEBUG nova.virt.hardware [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 940.540397] env[63202]: DEBUG nova.virt.hardware [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 940.540463] env[63202]: DEBUG nova.virt.hardware [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 940.541519] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85b71017-c5c2-4624-80aa-d356f893cb21 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.551615] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db74087f-6f0a-4c33-9a14-626c373070f3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.566411] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Instance VIF info [] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 940.572430] env[63202]: DEBUG oslo.service.loopingcall [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 940.576974] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 940.577409] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-50a81888-14c4-4ee8-a153-a5b3aef8e38b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.598849] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 940.598849] env[63202]: value = "task-1385555" [ 940.598849] env[63202]: _type = "Task" [ 940.598849] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.611305] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385555, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.631997] env[63202]: DEBUG nova.network.neutron [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Updating instance_info_cache with network_info: [{"id": "5eeb4ae8-9356-402a-965d-224788f04129", "address": "fa:16:3e:b1:37:8e", "network": {"id": "0aa55fee-953e-4c2d-b88a-43b0cf86c7f3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-878071414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7d28fea097541adadf0839940568409", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5eeb4ae8-93", "ovs_interfaceid": "5eeb4ae8-9356-402a-965d-224788f04129", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.763865] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b402e1a-6f96-4b6b-865b-b53f987430c5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.773158] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bd6b8c8-9a85-4d42-9b4d-5fc1cdb8ddc4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.809080] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6882237-c9a2-462d-be63-83d6315260f9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.817368] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef301764-96f2-49e4-bf2b-d69aac55a8dc {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.830418] env[63202]: DEBUG nova.compute.provider_tree [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Updating inventory in ProviderTree for provider 79b33d17-3e75-494c-a550-67b275de2079 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 940.854940] env[63202]: DEBUG nova.network.neutron [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Updating instance_info_cache with network_info: [{"id": "c7da6929-1bd1-4e2b-bcb8-609863cf6861", "address": "fa:16:3e:cf:6c:83", "network": {"id": "18390479-16ce-4012-9ba6-abf19b5616d1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1376026898-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22352c25bca8416a948014391a5389ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "085fb0ff-9285-4f1d-a008-a14da4844357", "external-id": "nsx-vlan-transportzone-729", "segmentation_id": 729, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7da6929-1b", "ovs_interfaceid": "c7da6929-1bd1-4e2b-bcb8-609863cf6861", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.859602] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385546, 'name': CreateVM_Task, 'duration_secs': 1.883848} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.860573] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 940.860685] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.860854] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.861437] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 940.863943] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f7eaf0ae-e3c0-4d9b-a36b-0eb1d51e57b5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.870736] env[63202]: DEBUG oslo_vmware.api [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Waiting for the task: (returnval){ [ 940.870736] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52ad8eb0-5337-0fff-eb16-fd6d71a2f72a" [ 940.870736] env[63202]: _type = "Task" [ 940.870736] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.882750] env[63202]: DEBUG oslo_vmware.api [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52ad8eb0-5337-0fff-eb16-fd6d71a2f72a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.885948] env[63202]: DEBUG oslo_vmware.api [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Task: {'id': task-1385552, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.936861] env[63202]: DEBUG nova.network.neutron [None req-8ba75055-34c9-4e5c-a6f6-df51e7ac604b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Updating instance_info_cache with network_info: [{"id": "e5621970-d0a8-4803-94ef-f6d577bdf6d2", "address": "fa:16:3e:00:ba:72", "network": {"id": "345538f4-e791-4fc6-9719-237c11f1382a", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-449651008-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.170", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a9e939cdf4a94a8aa0107a5761771c2c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa09e855-8af1-419b-b78d-8ffcc94b1bfb", "external-id": "nsx-vlan-transportzone-901", "segmentation_id": 901, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape5621970-d0", "ovs_interfaceid": "e5621970-d0a8-4803-94ef-f6d577bdf6d2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.943929] env[63202]: DEBUG oslo_vmware.api [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385553, 'name': ReconfigVM_Task, 'duration_secs': 0.282002} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.944446] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Reconfigured VM instance instance-0000004f to attach disk [datastore2] 8755bceb-d510-4429-bd98-d6a63faf739d/8755bceb-d510-4429-bd98-d6a63faf739d.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 940.945046] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-747546fb-b2c8-4e70-95c3-f126be24d748 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.952610] env[63202]: DEBUG oslo_vmware.api [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 940.952610] env[63202]: value = "task-1385556" [ 940.952610] env[63202]: _type = "Task" [ 940.952610] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.960346] env[63202]: DEBUG oslo_vmware.api [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385556, 'name': Rename_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.011909] env[63202]: DEBUG oslo_vmware.api [None req-02dae3ed-6d57-4066-a061-72500d73ac65 tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Task: {'id': task-1385554, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.305499} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.012280] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-02dae3ed-6d57-4066-a061-72500d73ac65 tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 941.012500] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-02dae3ed-6d57-4066-a061-72500d73ac65 tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Deleted contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 941.012736] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-02dae3ed-6d57-4066-a061-72500d73ac65 tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 941.012958] env[63202]: INFO nova.compute.manager [None req-02dae3ed-6d57-4066-a061-72500d73ac65 tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Took 3.25 seconds to destroy the instance on the hypervisor. [ 941.013256] env[63202]: DEBUG oslo.service.loopingcall [None req-02dae3ed-6d57-4066-a061-72500d73ac65 tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 941.013491] env[63202]: DEBUG nova.compute.manager [-] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 941.013616] env[63202]: DEBUG nova.network.neutron [-] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 941.109897] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385555, 'name': CreateVM_Task, 'duration_secs': 0.354581} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.109897] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 941.109897] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.134517] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Releasing lock "refresh_cache-7437595c-fa35-483e-95f3-b75405b6bd13" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.134846] env[63202]: DEBUG nova.compute.manager [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Instance network_info: |[{"id": "5eeb4ae8-9356-402a-965d-224788f04129", "address": "fa:16:3e:b1:37:8e", "network": {"id": "0aa55fee-953e-4c2d-b88a-43b0cf86c7f3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-878071414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7d28fea097541adadf0839940568409", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5eeb4ae8-93", "ovs_interfaceid": "5eeb4ae8-9356-402a-965d-224788f04129", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 941.136161] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b1:37:8e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5eeb4ae8-9356-402a-965d-224788f04129', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 941.144180] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Creating folder: Project (f7d28fea097541adadf0839940568409). Parent ref: group-v294090. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 941.144546] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ccf45958-4260-412e-a0a0-7222671f80d1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.154757] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Created folder: Project (f7d28fea097541adadf0839940568409) in parent group-v294090. [ 941.154957] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Creating folder: Instances. Parent ref: group-v294166. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 941.155224] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b55a01a9-7125-49b3-8d4f-c85baab47f21 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.163535] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Created folder: Instances in parent group-v294166. [ 941.163807] env[63202]: DEBUG oslo.service.loopingcall [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 941.164011] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 941.164226] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6fd63bf2-44a3-4082-abd6-a957ab71793b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.184170] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 941.184170] env[63202]: value = "task-1385559" [ 941.184170] env[63202]: _type = "Task" [ 941.184170] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.194244] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385559, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.254539] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 941.360904] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Releasing lock "refresh_cache-e775e5e1-521a-4fc7-80e6-bcb6a70516c5" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.376201] env[63202]: DEBUG nova.scheduler.client.report [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Updated inventory for provider 79b33d17-3e75-494c-a550-67b275de2079 with generation 94 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 941.376549] env[63202]: DEBUG nova.compute.provider_tree [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Updating resource provider 79b33d17-3e75-494c-a550-67b275de2079 generation from 94 to 95 during operation: update_inventory {{(pid=63202) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 941.376794] env[63202]: DEBUG nova.compute.provider_tree [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Updating inventory in ProviderTree for provider 79b33d17-3e75-494c-a550-67b275de2079 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 941.390013] env[63202]: DEBUG oslo_vmware.api [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Task: {'id': task-1385552, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.395354] env[63202]: DEBUG oslo_vmware.api [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52ad8eb0-5337-0fff-eb16-fd6d71a2f72a, 'name': SearchDatastore_Task, 'duration_secs': 0.031516} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.395675] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.395927] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 941.396177] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.396370] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.396578] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 941.396887] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.397232] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 941.397501] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-83146811-6303-4bef-bd97-2d22466c22a6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.402287] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a5fae828-5492-46b4-a11b-67eb40322357 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.406408] env[63202]: DEBUG oslo_vmware.api [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Waiting for the task: (returnval){ [ 941.406408] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52550d34-2cef-86e0-8da3-b1ab63a3ae68" [ 941.406408] env[63202]: _type = "Task" [ 941.406408] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.411192] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 941.411392] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 941.412475] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b755f4fa-a586-4e98-9346-7325359fac2b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.418116] env[63202]: DEBUG oslo_vmware.api [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52550d34-2cef-86e0-8da3-b1ab63a3ae68, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.421499] env[63202]: DEBUG oslo_vmware.api [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Waiting for the task: (returnval){ [ 941.421499] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52eeec4a-4448-914a-f32c-4c7bc5df11c2" [ 941.421499] env[63202]: _type = "Task" [ 941.421499] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.432275] env[63202]: DEBUG oslo_vmware.api [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52eeec4a-4448-914a-f32c-4c7bc5df11c2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.445335] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8ba75055-34c9-4e5c-a6f6-df51e7ac604b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Releasing lock "refresh_cache-0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.460725] env[63202]: DEBUG nova.compute.manager [req-fff07e0a-2644-460b-b8fa-8e2fded3d828 req-0569f9e1-34f6-4a9c-8633-45e7df09b0ae service nova] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Received event network-changed-5eeb4ae8-9356-402a-965d-224788f04129 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 941.460725] env[63202]: DEBUG nova.compute.manager [req-fff07e0a-2644-460b-b8fa-8e2fded3d828 req-0569f9e1-34f6-4a9c-8633-45e7df09b0ae service nova] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Refreshing instance network info cache due to event network-changed-5eeb4ae8-9356-402a-965d-224788f04129. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 941.460927] env[63202]: DEBUG oslo_concurrency.lockutils [req-fff07e0a-2644-460b-b8fa-8e2fded3d828 req-0569f9e1-34f6-4a9c-8633-45e7df09b0ae service nova] Acquiring lock "refresh_cache-7437595c-fa35-483e-95f3-b75405b6bd13" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.463510] env[63202]: DEBUG oslo_concurrency.lockutils [req-fff07e0a-2644-460b-b8fa-8e2fded3d828 req-0569f9e1-34f6-4a9c-8633-45e7df09b0ae service nova] Acquired lock "refresh_cache-7437595c-fa35-483e-95f3-b75405b6bd13" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.463510] env[63202]: DEBUG nova.network.neutron [req-fff07e0a-2644-460b-b8fa-8e2fded3d828 req-0569f9e1-34f6-4a9c-8633-45e7df09b0ae service nova] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Refreshing network info cache for port 5eeb4ae8-9356-402a-965d-224788f04129 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 941.468509] env[63202]: DEBUG oslo_vmware.api [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385556, 'name': Rename_Task, 'duration_secs': 0.128713} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.469258] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 941.469258] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4323323d-6c38-4c67-bf5f-65908db0fea0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.475279] env[63202]: DEBUG oslo_vmware.api [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 941.475279] env[63202]: value = "task-1385560" [ 941.475279] env[63202]: _type = "Task" [ 941.475279] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.483581] env[63202]: DEBUG oslo_vmware.api [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385560, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.699650] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385559, 'name': CreateVM_Task, 'duration_secs': 0.332542} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.699821] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 941.700536] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.883359] env[63202]: DEBUG oslo_vmware.api [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Task: {'id': task-1385552, 'name': PowerOnVM_Task, 'duration_secs': 1.435495} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.883674] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 941.883892] env[63202]: INFO nova.compute.manager [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Took 11.11 seconds to spawn the instance on the hypervisor. [ 941.884202] env[63202]: DEBUG nova.compute.manager [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 941.885040] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.436s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.885686] env[63202]: DEBUG nova.compute.manager [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 941.889301] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-feea54a5-aa43-439d-9ee9-d13a4969221c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.896106] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d993524b-f10f-4ca3-a419-4ae825cd0a96 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.280s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.896280] env[63202]: DEBUG nova.objects.instance [None req-d993524b-f10f-4ca3-a419-4ae825cd0a96 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Lazy-loading 'resources' on Instance uuid 61a6528e-4e4e-49b7-be7b-dd30bd8023d8 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 941.917503] env[63202]: DEBUG oslo_vmware.api [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52550d34-2cef-86e0-8da3-b1ab63a3ae68, 'name': SearchDatastore_Task, 'duration_secs': 0.01657} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.918579] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.918939] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 941.919335] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.919867] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.920360] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 941.921491] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d5e86291-01e8-4b48-b9c1-9ec154e3d8ac {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.931410] env[63202]: DEBUG oslo_vmware.api [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 941.931410] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52fdfe28-78d8-05c8-a655-89f4e757fb00" [ 941.931410] env[63202]: _type = "Task" [ 941.931410] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.936394] env[63202]: DEBUG oslo_vmware.api [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52eeec4a-4448-914a-f32c-4c7bc5df11c2, 'name': SearchDatastore_Task, 'duration_secs': 0.025972} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.940384] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d1c4200-3f31-45ee-9370-7d487094df26 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.949827] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ba75055-34c9-4e5c-a6f6-df51e7ac604b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 941.949827] env[63202]: DEBUG oslo_vmware.api [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52fdfe28-78d8-05c8-a655-89f4e757fb00, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.951297] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3c53b927-7f4d-42ee-b472-44cc43c7cb0f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.952762] env[63202]: DEBUG oslo_vmware.api [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Waiting for the task: (returnval){ [ 941.952762] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]5230c4e9-7e5d-8ab0-16eb-29b6feddad32" [ 941.952762] env[63202]: _type = "Task" [ 941.952762] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.958334] env[63202]: DEBUG oslo_vmware.api [None req-8ba75055-34c9-4e5c-a6f6-df51e7ac604b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Waiting for the task: (returnval){ [ 941.958334] env[63202]: value = "task-1385561" [ 941.958334] env[63202]: _type = "Task" [ 941.958334] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.965621] env[63202]: DEBUG oslo_vmware.api [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]5230c4e9-7e5d-8ab0-16eb-29b6feddad32, 'name': SearchDatastore_Task, 'duration_secs': 0.009576} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.967676] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.967965] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore2] f3128c09-3680-4b0b-b463-3d6cd203fcf4/f3128c09-3680-4b0b-b463-3d6cd203fcf4.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 941.969024] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.969258] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 941.969490] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d84a387b-2688-48d4-a26f-369a5e9eb630 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.972040] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-95bc62df-b82b-4c36-ba1e-4796d4f0f138 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.977449] env[63202]: DEBUG oslo_vmware.api [None req-8ba75055-34c9-4e5c-a6f6-df51e7ac604b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385561, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.982152] env[63202]: DEBUG oslo_vmware.api [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Waiting for the task: (returnval){ [ 941.982152] env[63202]: value = "task-1385562" [ 941.982152] env[63202]: _type = "Task" [ 941.982152] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.989975] env[63202]: DEBUG oslo_vmware.api [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385560, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.990348] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 941.990524] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 941.991864] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d601c0a-7344-418f-98f4-8e707dd7ead5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.997065] env[63202]: DEBUG oslo_vmware.api [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Task: {'id': task-1385562, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.000273] env[63202]: DEBUG oslo_vmware.api [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Waiting for the task: (returnval){ [ 942.000273] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]520b5484-0b83-ebfa-9976-7a088dac6405" [ 942.000273] env[63202]: _type = "Task" [ 942.000273] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.009932] env[63202]: DEBUG oslo_vmware.api [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]520b5484-0b83-ebfa-9976-7a088dac6405, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.178029] env[63202]: DEBUG nova.network.neutron [req-fff07e0a-2644-460b-b8fa-8e2fded3d828 req-0569f9e1-34f6-4a9c-8633-45e7df09b0ae service nova] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Updated VIF entry in instance network info cache for port 5eeb4ae8-9356-402a-965d-224788f04129. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 942.178331] env[63202]: DEBUG nova.network.neutron [req-fff07e0a-2644-460b-b8fa-8e2fded3d828 req-0569f9e1-34f6-4a9c-8633-45e7df09b0ae service nova] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Updating instance_info_cache with network_info: [{"id": "5eeb4ae8-9356-402a-965d-224788f04129", "address": "fa:16:3e:b1:37:8e", "network": {"id": "0aa55fee-953e-4c2d-b88a-43b0cf86c7f3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-878071414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7d28fea097541adadf0839940568409", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5eeb4ae8-93", "ovs_interfaceid": "5eeb4ae8-9356-402a-965d-224788f04129", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.179865] env[63202]: DEBUG nova.network.neutron [-] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.248803] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 942.253291] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 942.253541] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63202) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 942.397432] env[63202]: DEBUG nova.compute.utils [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 942.399378] env[63202]: DEBUG nova.compute.manager [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 942.399497] env[63202]: DEBUG nova.network.neutron [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 942.425870] env[63202]: INFO nova.compute.manager [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Took 32.53 seconds to build instance. [ 942.450613] env[63202]: DEBUG oslo_vmware.api [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52fdfe28-78d8-05c8-a655-89f4e757fb00, 'name': SearchDatastore_Task, 'duration_secs': 0.019655} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.451150] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.451261] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 942.452721] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.454762] env[63202]: DEBUG nova.policy [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2b9d1001300c476a896f335b3acf041b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f651d18a9559499da27f301ee75bb2fb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 942.473115] env[63202]: DEBUG oslo_vmware.api [None req-8ba75055-34c9-4e5c-a6f6-df51e7ac604b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385561, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.494130] env[63202]: DEBUG oslo_vmware.api [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385560, 'name': PowerOnVM_Task, 'duration_secs': 0.540341} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.498840] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 942.498840] env[63202]: INFO nova.compute.manager [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Took 9.28 seconds to spawn the instance on the hypervisor. [ 942.498840] env[63202]: DEBUG nova.compute.manager [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 942.498840] env[63202]: DEBUG oslo_vmware.api [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Task: {'id': task-1385562, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.503459] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f24c9c68-e87e-4a59-896e-c4852b90ee5d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.518034] env[63202]: DEBUG oslo_vmware.api [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]520b5484-0b83-ebfa-9976-7a088dac6405, 'name': SearchDatastore_Task, 'duration_secs': 0.009792} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.522105] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dca2a8dc-0bef-4f44-82e0-8f2460e64a41 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.528515] env[63202]: DEBUG oslo_vmware.api [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Waiting for the task: (returnval){ [ 942.528515] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52469619-8cce-9fde-ecd2-d2cb2c31c269" [ 942.528515] env[63202]: _type = "Task" [ 942.528515] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.543867] env[63202]: DEBUG oslo_vmware.api [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52469619-8cce-9fde-ecd2-d2cb2c31c269, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.681596] env[63202]: DEBUG oslo_concurrency.lockutils [req-fff07e0a-2644-460b-b8fa-8e2fded3d828 req-0569f9e1-34f6-4a9c-8633-45e7df09b0ae service nova] Releasing lock "refresh_cache-7437595c-fa35-483e-95f3-b75405b6bd13" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.685680] env[63202]: INFO nova.compute.manager [-] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Took 1.67 seconds to deallocate network for instance. [ 942.771289] env[63202]: DEBUG nova.network.neutron [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Successfully created port: 836478e8-9299-44c3-a4b0-957cd1706387 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 942.784823] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc751a7a-6eb2-4cc7-9d09-c47f88c78d21 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.793434] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99c7d77f-d98c-4661-862d-7f5d2bffc7c8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.829151] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c48ca04-c807-440c-a495-4a7ba105459e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.840430] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36dbb4aa-06d7-4033-b6fb-67f8c728cf45 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.857116] env[63202]: DEBUG nova.compute.provider_tree [None req-d993524b-f10f-4ca3-a419-4ae825cd0a96 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 942.876688] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3859903-2a37-4276-b588-9bd598f9ad17 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.898015] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Updating instance 'e775e5e1-521a-4fc7-80e6-bcb6a70516c5' progress to 0 {{(pid=63202) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 942.909520] env[63202]: DEBUG nova.compute.manager [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 942.930021] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b9272792-9c96-4fbc-a625-47aa2dbd16f8 tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Lock "d0d6e380-9337-4f69-8434-6a33ac8a33d6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 84.405s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.972413] env[63202]: DEBUG oslo_vmware.api [None req-8ba75055-34c9-4e5c-a6f6-df51e7ac604b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385561, 'name': PowerOnVM_Task, 'duration_secs': 0.587661} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.972531] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ba75055-34c9-4e5c-a6f6-df51e7ac604b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 942.973612] env[63202]: DEBUG nova.compute.manager [None req-8ba75055-34c9-4e5c-a6f6-df51e7ac604b tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 942.973978] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ba67338-06cd-46bb-9501-0902e546af87 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.994533] env[63202]: DEBUG oslo_vmware.api [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Task: {'id': task-1385562, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.029023] env[63202]: INFO nova.compute.manager [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Took 31.18 seconds to build instance. [ 943.042195] env[63202]: DEBUG oslo_vmware.api [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52469619-8cce-9fde-ecd2-d2cb2c31c269, 'name': SearchDatastore_Task, 'duration_secs': 0.051676} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.042330] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.042528] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore2] fc7d3924-a624-4ea4-890c-7628595fb733/fc7d3924-a624-4ea4-890c-7628595fb733.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 943.042781] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.042982] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 943.043205] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e70383d6-f2e4-4b46-8ec7-f68f4339b499 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.045181] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-62815fe3-c5af-47da-a83d-e9ce5199c3eb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.052118] env[63202]: DEBUG oslo_vmware.api [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Waiting for the task: (returnval){ [ 943.052118] env[63202]: value = "task-1385563" [ 943.052118] env[63202]: _type = "Task" [ 943.052118] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.056671] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 943.056863] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 943.057878] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1eeb1f3c-6379-4c37-98c0-3cb84fc66747 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.065921] env[63202]: DEBUG oslo_vmware.api [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': task-1385563, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.068844] env[63202]: DEBUG oslo_vmware.api [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 943.068844] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]524c7093-33fe-e898-ca42-bfc66a4f06b7" [ 943.068844] env[63202]: _type = "Task" [ 943.068844] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.076036] env[63202]: DEBUG oslo_vmware.api [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]524c7093-33fe-e898-ca42-bfc66a4f06b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.880954] env[63202]: DEBUG oslo_concurrency.lockutils [None req-02dae3ed-6d57-4066-a061-72500d73ac65 tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.881386] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager.update_available_resource {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 943.882421] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d316d71b-590a-440a-b0fc-2e046c2315ac tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Acquiring lock "d0d6e380-9337-4f69-8434-6a33ac8a33d6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.882643] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d316d71b-590a-440a-b0fc-2e046c2315ac tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Lock "d0d6e380-9337-4f69-8434-6a33ac8a33d6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.882838] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d316d71b-590a-440a-b0fc-2e046c2315ac tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Acquiring lock "d0d6e380-9337-4f69-8434-6a33ac8a33d6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.883029] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d316d71b-590a-440a-b0fc-2e046c2315ac tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Lock "d0d6e380-9337-4f69-8434-6a33ac8a33d6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.883203] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d316d71b-590a-440a-b0fc-2e046c2315ac tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Lock "d0d6e380-9337-4f69-8434-6a33ac8a33d6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.885027] env[63202]: DEBUG nova.scheduler.client.report [None req-d993524b-f10f-4ca3-a419-4ae825cd0a96 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 943.888904] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 943.893543] env[63202]: DEBUG oslo_concurrency.lockutils [None req-6565c11f-66c4-43f6-988b-43f13d0f5e9b tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "8755bceb-d510-4429-bd98-d6a63faf739d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.999s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.894600] env[63202]: DEBUG nova.compute.manager [req-5926ef6e-6283-42df-860d-70fb9e98619e req-190f2293-16f9-4348-b29a-8392bd9f9988 service nova] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Received event network-vif-deleted-a0175615-5853-4224-8504-2a296446e39b {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 943.895327] env[63202]: INFO nova.compute.manager [None req-d316d71b-590a-440a-b0fc-2e046c2315ac tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Terminating instance [ 943.897450] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-26cdcca9-0a20-491a-8d05-b13c3d8a5580 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.907548] env[63202]: DEBUG nova.compute.manager [None req-d316d71b-590a-440a-b0fc-2e046c2315ac tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 943.907818] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d316d71b-590a-440a-b0fc-2e046c2315ac tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 943.908833] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5db8236a-9cae-49d0-983a-09c780ded4d5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.918939] env[63202]: DEBUG oslo_vmware.api [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]524c7093-33fe-e898-ca42-bfc66a4f06b7, 'name': SearchDatastore_Task, 'duration_secs': 0.034206} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.926749] env[63202]: DEBUG oslo_vmware.api [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for the task: (returnval){ [ 943.926749] env[63202]: value = "task-1385564" [ 943.926749] env[63202]: _type = "Task" [ 943.926749] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.927209] env[63202]: DEBUG oslo_vmware.api [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Task: {'id': task-1385562, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.695495} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.927418] env[63202]: DEBUG oslo_vmware.api [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': task-1385563, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.928016] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-94317e58-397e-413c-9ae0-2e987777517d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.930305] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore2] f3128c09-3680-4b0b-b463-3d6cd203fcf4/f3128c09-3680-4b0b-b463-3d6cd203fcf4.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 943.930532] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 943.935710] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0b80ca81-27e6-48a1-a56e-9a293fa763b0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.937438] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-d316d71b-590a-440a-b0fc-2e046c2315ac tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 943.938238] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2fa215e3-3d4c-481b-bc4c-02cb9171e762 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.941740] env[63202]: DEBUG oslo_vmware.api [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 943.941740] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]5254edb4-22d9-2e33-6279-7885024f13ca" [ 943.941740] env[63202]: _type = "Task" [ 943.941740] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.949018] env[63202]: DEBUG oslo_vmware.api [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385564, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.950319] env[63202]: DEBUG oslo_vmware.api [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Waiting for the task: (returnval){ [ 943.950319] env[63202]: value = "task-1385565" [ 943.950319] env[63202]: _type = "Task" [ 943.950319] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.950986] env[63202]: DEBUG oslo_vmware.api [None req-d316d71b-590a-440a-b0fc-2e046c2315ac tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Waiting for the task: (returnval){ [ 943.950986] env[63202]: value = "task-1385566" [ 943.950986] env[63202]: _type = "Task" [ 943.950986] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.962937] env[63202]: DEBUG oslo_vmware.api [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]5254edb4-22d9-2e33-6279-7885024f13ca, 'name': SearchDatastore_Task, 'duration_secs': 0.011407} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.963621] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.963908] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore2] 7437595c-fa35-483e-95f3-b75405b6bd13/7437595c-fa35-483e-95f3-b75405b6bd13.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 943.964524] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2b7f3fd7-5639-46c9-a83a-d0d67286e3e8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.972825] env[63202]: DEBUG oslo_vmware.api [None req-d316d71b-590a-440a-b0fc-2e046c2315ac tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Task: {'id': task-1385566, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.973051] env[63202]: DEBUG oslo_vmware.api [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Task: {'id': task-1385565, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.978888] env[63202]: DEBUG oslo_vmware.api [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 943.978888] env[63202]: value = "task-1385567" [ 943.978888] env[63202]: _type = "Task" [ 943.978888] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.988353] env[63202]: DEBUG oslo_vmware.api [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385567, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.172699] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4e400d69-10f2-4327-9888-f1914bbefb0f tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "8755bceb-d510-4429-bd98-d6a63faf739d" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.172888] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4e400d69-10f2-4327-9888-f1914bbefb0f tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "8755bceb-d510-4429-bd98-d6a63faf739d" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.358780] env[63202]: DEBUG nova.network.neutron [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Successfully updated port: 836478e8-9299-44c3-a4b0-957cd1706387 {{(pid=63202) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 944.403032] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d993524b-f10f-4ca3-a419-4ae825cd0a96 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.507s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.405386] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.406403] env[63202]: DEBUG nova.compute.manager [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 944.409779] env[63202]: DEBUG oslo_concurrency.lockutils [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.570s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.410601] env[63202]: INFO nova.compute.claims [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 944.423506] env[63202]: DEBUG oslo_vmware.api [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': task-1385563, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.350543} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.423909] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore2] fc7d3924-a624-4ea4-890c-7628595fb733/fc7d3924-a624-4ea4-890c-7628595fb733.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 944.424555] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 944.424647] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-17c4efdc-4235-4f17-9009-b400d817ff6b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.436604] env[63202]: DEBUG nova.virt.hardware [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 944.436869] env[63202]: DEBUG nova.virt.hardware [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 944.437038] env[63202]: DEBUG nova.virt.hardware [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 944.437227] env[63202]: DEBUG nova.virt.hardware [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 944.437374] env[63202]: DEBUG nova.virt.hardware [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 944.437519] env[63202]: DEBUG nova.virt.hardware [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 944.437751] env[63202]: DEBUG nova.virt.hardware [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 944.437914] env[63202]: DEBUG nova.virt.hardware [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 944.438116] env[63202]: DEBUG nova.virt.hardware [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 944.438314] env[63202]: DEBUG nova.virt.hardware [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 944.438795] env[63202]: DEBUG nova.virt.hardware [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 944.440839] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44d0be62-a92c-417b-b8d6-fd284c2cd187 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.443985] env[63202]: DEBUG oslo_vmware.api [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Waiting for the task: (returnval){ [ 944.443985] env[63202]: value = "task-1385568" [ 944.443985] env[63202]: _type = "Task" [ 944.443985] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.444446] env[63202]: INFO nova.scheduler.client.report [None req-d993524b-f10f-4ca3-a419-4ae825cd0a96 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Deleted allocations for instance 61a6528e-4e4e-49b7-be7b-dd30bd8023d8 [ 944.457539] env[63202]: DEBUG oslo_vmware.api [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385564, 'name': PowerOffVM_Task, 'duration_secs': 0.3169} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.463062] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b70ff228-3d82-4fa0-99ec-28e3a51e3edf {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.469611] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 944.469800] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Updating instance 'e775e5e1-521a-4fc7-80e6-bcb6a70516c5' progress to 17 {{(pid=63202) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 944.476424] env[63202]: DEBUG oslo_vmware.api [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': task-1385568, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.491950] env[63202]: DEBUG oslo_vmware.api [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Task: {'id': task-1385565, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071601} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.492989] env[63202]: DEBUG oslo_vmware.api [None req-d316d71b-590a-440a-b0fc-2e046c2315ac tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Task: {'id': task-1385566, 'name': PowerOffVM_Task} progress is 100%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.496310] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 944.497339] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91fb9f30-7478-4f78-a12b-0ca917a9fefc {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.504866] env[63202]: DEBUG oslo_vmware.api [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385567, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.522405] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Reconfiguring VM instance instance-00000050 to attach disk [datastore2] f3128c09-3680-4b0b-b463-3d6cd203fcf4/f3128c09-3680-4b0b-b463-3d6cd203fcf4.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 944.523417] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7870bca7-968d-4b12-9fc5-1a3101087f06 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.542512] env[63202]: DEBUG oslo_vmware.api [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Waiting for the task: (returnval){ [ 944.542512] env[63202]: value = "task-1385569" [ 944.542512] env[63202]: _type = "Task" [ 944.542512] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.550792] env[63202]: DEBUG oslo_vmware.api [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Task: {'id': task-1385569, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.677418] env[63202]: DEBUG nova.compute.utils [None req-4e400d69-10f2-4327-9888-f1914bbefb0f tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 944.862032] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Acquiring lock "refresh_cache-a961b2a0-39c1-4267-9229-068e2b6ecd67" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.862206] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Acquired lock "refresh_cache-a961b2a0-39c1-4267-9229-068e2b6ecd67" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.862394] env[63202]: DEBUG nova.network.neutron [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 944.958827] env[63202]: DEBUG oslo_vmware.api [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': task-1385568, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.962424] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d993524b-f10f-4ca3-a419-4ae825cd0a96 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Lock "61a6528e-4e4e-49b7-be7b-dd30bd8023d8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.737s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.968934] env[63202]: DEBUG oslo_vmware.api [None req-d316d71b-590a-440a-b0fc-2e046c2315ac tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Task: {'id': task-1385566, 'name': PowerOffVM_Task, 'duration_secs': 0.579381} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.969215] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-d316d71b-590a-440a-b0fc-2e046c2315ac tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 944.969384] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d316d71b-590a-440a-b0fc-2e046c2315ac tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 944.969629] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c5f62c87-7ec4-42c7-b31b-d8549531a085 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.979287] env[63202]: DEBUG nova.virt.hardware [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 944.979523] env[63202]: DEBUG nova.virt.hardware [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 944.979685] env[63202]: DEBUG nova.virt.hardware [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 944.979860] env[63202]: DEBUG nova.virt.hardware [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 944.979996] env[63202]: DEBUG nova.virt.hardware [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 944.980167] env[63202]: DEBUG nova.virt.hardware [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 944.980394] env[63202]: DEBUG nova.virt.hardware [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 944.980557] env[63202]: DEBUG nova.virt.hardware [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 944.980721] env[63202]: DEBUG nova.virt.hardware [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 944.980882] env[63202]: DEBUG nova.virt.hardware [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 944.981065] env[63202]: DEBUG nova.virt.hardware [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 944.986409] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5a683a51-2143-44d6-8065-7c3d9964bf21 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.005878] env[63202]: DEBUG oslo_vmware.api [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385567, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.899186} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.007300] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore2] 7437595c-fa35-483e-95f3-b75405b6bd13/7437595c-fa35-483e-95f3-b75405b6bd13.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 945.007560] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 945.007899] env[63202]: DEBUG oslo_vmware.api [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for the task: (returnval){ [ 945.007899] env[63202]: value = "task-1385571" [ 945.007899] env[63202]: _type = "Task" [ 945.007899] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.008135] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6fd7e043-dfb5-4595-9c0d-f284b79138ab {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.019585] env[63202]: DEBUG oslo_vmware.api [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385571, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.021298] env[63202]: DEBUG oslo_vmware.api [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 945.021298] env[63202]: value = "task-1385572" [ 945.021298] env[63202]: _type = "Task" [ 945.021298] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.034018] env[63202]: DEBUG oslo_vmware.api [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385572, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.040831] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d316d71b-590a-440a-b0fc-2e046c2315ac tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 945.041062] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d316d71b-590a-440a-b0fc-2e046c2315ac tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Deleting contents of the VM from datastore datastore2 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 945.041270] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-d316d71b-590a-440a-b0fc-2e046c2315ac tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Deleting the datastore file [datastore2] d0d6e380-9337-4f69-8434-6a33ac8a33d6 {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 945.041632] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cf6d7229-a3c5-4b0b-8b41-124c68b929aa {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.052602] env[63202]: DEBUG oslo_vmware.api [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Task: {'id': task-1385569, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.054111] env[63202]: DEBUG oslo_vmware.api [None req-d316d71b-590a-440a-b0fc-2e046c2315ac tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Waiting for the task: (returnval){ [ 945.054111] env[63202]: value = "task-1385573" [ 945.054111] env[63202]: _type = "Task" [ 945.054111] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.063501] env[63202]: DEBUG oslo_vmware.api [None req-d316d71b-590a-440a-b0fc-2e046c2315ac tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Task: {'id': task-1385573, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.184227] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4e400d69-10f2-4327-9888-f1914bbefb0f tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "8755bceb-d510-4429-bd98-d6a63faf739d" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.011s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.395556] env[63202]: DEBUG nova.network.neutron [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 945.460320] env[63202]: DEBUG oslo_vmware.api [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': task-1385568, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.683744} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.463048] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 945.463963] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8e377b4-0775-44a9-a239-f77734b9a0ab {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.487748] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] fc7d3924-a624-4ea4-890c-7628595fb733/fc7d3924-a624-4ea4-890c-7628595fb733.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 945.490732] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dd0fe4c6-ff8f-4df8-b906-65e24cd6199c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.515311] env[63202]: DEBUG oslo_vmware.api [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Waiting for the task: (returnval){ [ 945.515311] env[63202]: value = "task-1385574" [ 945.515311] env[63202]: _type = "Task" [ 945.515311] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.525129] env[63202]: DEBUG oslo_vmware.api [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385571, 'name': ReconfigVM_Task, 'duration_secs': 0.364702} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.528223] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Updating instance 'e775e5e1-521a-4fc7-80e6-bcb6a70516c5' progress to 33 {{(pid=63202) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 945.539080] env[63202]: DEBUG oslo_vmware.api [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': task-1385574, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.541637] env[63202]: DEBUG nova.network.neutron [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Updating instance_info_cache with network_info: [{"id": "836478e8-9299-44c3-a4b0-957cd1706387", "address": "fa:16:3e:b0:50:ca", "network": {"id": "69295efd-3c50-45e5-bc09-4f0cef6b8f0e", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1062402518-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f651d18a9559499da27f301ee75bb2fb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dced2f3d-7fd3-4a42-836d-9f02dab4c949", "external-id": "nsx-vlan-transportzone-117", "segmentation_id": 117, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap836478e8-92", "ovs_interfaceid": "836478e8-9299-44c3-a4b0-957cd1706387", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.546147] env[63202]: DEBUG oslo_vmware.api [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385572, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076584} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.549055] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 945.549574] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3dd9289-e2ed-4c82-948c-667127b60cfb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.570876] env[63202]: DEBUG oslo_vmware.api [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Task: {'id': task-1385569, 'name': ReconfigVM_Task, 'duration_secs': 0.572016} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.579835] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] 7437595c-fa35-483e-95f3-b75405b6bd13/7437595c-fa35-483e-95f3-b75405b6bd13.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 945.583603] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Reconfigured VM instance instance-00000050 to attach disk [datastore2] f3128c09-3680-4b0b-b463-3d6cd203fcf4/f3128c09-3680-4b0b-b463-3d6cd203fcf4.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 945.584262] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-36f47aa1-2647-42e4-9ade-1313dcfe14db {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.601432] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-299852dc-743d-4fa2-ab14-a72bca2ac7b4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.603784] env[63202]: DEBUG oslo_vmware.api [None req-d316d71b-590a-440a-b0fc-2e046c2315ac tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Task: {'id': task-1385573, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.17488} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.604449] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-d316d71b-590a-440a-b0fc-2e046c2315ac tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 945.604633] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d316d71b-590a-440a-b0fc-2e046c2315ac tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Deleted contents of the VM from datastore datastore2 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 945.606095] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d316d71b-590a-440a-b0fc-2e046c2315ac tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 945.606095] env[63202]: INFO nova.compute.manager [None req-d316d71b-590a-440a-b0fc-2e046c2315ac tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Took 1.70 seconds to destroy the instance on the hypervisor. [ 945.606095] env[63202]: DEBUG oslo.service.loopingcall [None req-d316d71b-590a-440a-b0fc-2e046c2315ac tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 945.606095] env[63202]: DEBUG nova.compute.manager [-] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 945.606095] env[63202]: DEBUG nova.network.neutron [-] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 945.609163] env[63202]: DEBUG oslo_vmware.api [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 945.609163] env[63202]: value = "task-1385575" [ 945.609163] env[63202]: _type = "Task" [ 945.609163] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.610448] env[63202]: DEBUG oslo_vmware.api [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Waiting for the task: (returnval){ [ 945.610448] env[63202]: value = "task-1385576" [ 945.610448] env[63202]: _type = "Task" [ 945.610448] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.618531] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7de759cc-c39d-4058-8625-09664c11b490 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Acquiring lock "3d1518d3-bed2-4b7b-af1d-d7da49e92874" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.618764] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7de759cc-c39d-4058-8625-09664c11b490 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Lock "3d1518d3-bed2-4b7b-af1d-d7da49e92874" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.618968] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7de759cc-c39d-4058-8625-09664c11b490 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Acquiring lock "3d1518d3-bed2-4b7b-af1d-d7da49e92874-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.619161] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7de759cc-c39d-4058-8625-09664c11b490 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Lock "3d1518d3-bed2-4b7b-af1d-d7da49e92874-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.619329] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7de759cc-c39d-4058-8625-09664c11b490 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Lock "3d1518d3-bed2-4b7b-af1d-d7da49e92874-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.621848] env[63202]: INFO nova.compute.manager [None req-7de759cc-c39d-4058-8625-09664c11b490 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Terminating instance [ 945.623836] env[63202]: DEBUG nova.compute.manager [None req-7de759cc-c39d-4058-8625-09664c11b490 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 945.624049] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-7de759cc-c39d-4058-8625-09664c11b490 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 945.627761] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e92b76af-88c8-48bb-8397-e380adbc4d8a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.630412] env[63202]: DEBUG oslo_vmware.api [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385575, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.634436] env[63202]: DEBUG oslo_vmware.api [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Task: {'id': task-1385576, 'name': Rename_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.641330] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-7de759cc-c39d-4058-8625-09664c11b490 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 945.641630] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2304f3d7-92ed-40c0-bc0f-9f403a4aa16d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.647948] env[63202]: DEBUG oslo_vmware.api [None req-7de759cc-c39d-4058-8625-09664c11b490 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Waiting for the task: (returnval){ [ 945.647948] env[63202]: value = "task-1385577" [ 945.647948] env[63202]: _type = "Task" [ 945.647948] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.659215] env[63202]: DEBUG oslo_vmware.api [None req-7de759cc-c39d-4058-8625-09664c11b490 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385577, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.677078] env[63202]: DEBUG nova.compute.manager [req-b92b834c-45c1-43a5-9bf0-65789a3827b5 req-3c4c7877-569f-47c5-96a4-4933f2d51a54 service nova] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Received event network-vif-plugged-836478e8-9299-44c3-a4b0-957cd1706387 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 945.677078] env[63202]: DEBUG oslo_concurrency.lockutils [req-b92b834c-45c1-43a5-9bf0-65789a3827b5 req-3c4c7877-569f-47c5-96a4-4933f2d51a54 service nova] Acquiring lock "a961b2a0-39c1-4267-9229-068e2b6ecd67-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.678665] env[63202]: DEBUG oslo_concurrency.lockutils [req-b92b834c-45c1-43a5-9bf0-65789a3827b5 req-3c4c7877-569f-47c5-96a4-4933f2d51a54 service nova] Lock "a961b2a0-39c1-4267-9229-068e2b6ecd67-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.678665] env[63202]: DEBUG oslo_concurrency.lockutils [req-b92b834c-45c1-43a5-9bf0-65789a3827b5 req-3c4c7877-569f-47c5-96a4-4933f2d51a54 service nova] Lock "a961b2a0-39c1-4267-9229-068e2b6ecd67-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.678665] env[63202]: DEBUG nova.compute.manager [req-b92b834c-45c1-43a5-9bf0-65789a3827b5 req-3c4c7877-569f-47c5-96a4-4933f2d51a54 service nova] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] No waiting events found dispatching network-vif-plugged-836478e8-9299-44c3-a4b0-957cd1706387 {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 945.678665] env[63202]: WARNING nova.compute.manager [req-b92b834c-45c1-43a5-9bf0-65789a3827b5 req-3c4c7877-569f-47c5-96a4-4933f2d51a54 service nova] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Received unexpected event network-vif-plugged-836478e8-9299-44c3-a4b0-957cd1706387 for instance with vm_state building and task_state spawning. [ 945.678665] env[63202]: DEBUG nova.compute.manager [req-b92b834c-45c1-43a5-9bf0-65789a3827b5 req-3c4c7877-569f-47c5-96a4-4933f2d51a54 service nova] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Received event network-changed-836478e8-9299-44c3-a4b0-957cd1706387 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 945.678885] env[63202]: DEBUG nova.compute.manager [req-b92b834c-45c1-43a5-9bf0-65789a3827b5 req-3c4c7877-569f-47c5-96a4-4933f2d51a54 service nova] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Refreshing instance network info cache due to event network-changed-836478e8-9299-44c3-a4b0-957cd1706387. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 945.678885] env[63202]: DEBUG oslo_concurrency.lockutils [req-b92b834c-45c1-43a5-9bf0-65789a3827b5 req-3c4c7877-569f-47c5-96a4-4933f2d51a54 service nova] Acquiring lock "refresh_cache-a961b2a0-39c1-4267-9229-068e2b6ecd67" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.782707] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c937eebe-8df9-4f7d-8633-a08df192f563 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.791512] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48332a1c-6c46-401b-a67c-c1b25aefed91 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.826910] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fafacff8-73dd-44bd-9625-e418f494ad78 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.834745] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bd961e1-e2fc-4be7-9ac0-57236fda72ef {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.853468] env[63202]: DEBUG nova.compute.provider_tree [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 946.025953] env[63202]: DEBUG oslo_vmware.api [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': task-1385574, 'name': ReconfigVM_Task, 'duration_secs': 0.343059} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.026386] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Reconfigured VM instance instance-0000004d to attach disk [datastore2] fc7d3924-a624-4ea4-890c-7628595fb733/fc7d3924-a624-4ea4-890c-7628595fb733.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 946.027011] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c3e9c19b-2b50-4d97-866d-94ecc447ad14 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.033809] env[63202]: DEBUG oslo_vmware.api [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Waiting for the task: (returnval){ [ 946.033809] env[63202]: value = "task-1385578" [ 946.033809] env[63202]: _type = "Task" [ 946.033809] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.043879] env[63202]: DEBUG nova.virt.hardware [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 946.044169] env[63202]: DEBUG nova.virt.hardware [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 946.044364] env[63202]: DEBUG nova.virt.hardware [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 946.044594] env[63202]: DEBUG nova.virt.hardware [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 946.044776] env[63202]: DEBUG nova.virt.hardware [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 946.044936] env[63202]: DEBUG nova.virt.hardware [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 946.045157] env[63202]: DEBUG nova.virt.hardware [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 946.045318] env[63202]: DEBUG nova.virt.hardware [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 946.045487] env[63202]: DEBUG nova.virt.hardware [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 946.045651] env[63202]: DEBUG nova.virt.hardware [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 946.045827] env[63202]: DEBUG nova.virt.hardware [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 946.051713] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Reconfiguring VM instance instance-00000048 to detach disk 2000 {{(pid=63202) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 946.052053] env[63202]: DEBUG oslo_vmware.api [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': task-1385578, 'name': Rename_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.052698] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Releasing lock "refresh_cache-a961b2a0-39c1-4267-9229-068e2b6ecd67" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.052848] env[63202]: DEBUG nova.compute.manager [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Instance network_info: |[{"id": "836478e8-9299-44c3-a4b0-957cd1706387", "address": "fa:16:3e:b0:50:ca", "network": {"id": "69295efd-3c50-45e5-bc09-4f0cef6b8f0e", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1062402518-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f651d18a9559499da27f301ee75bb2fb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dced2f3d-7fd3-4a42-836d-9f02dab4c949", "external-id": "nsx-vlan-transportzone-117", "segmentation_id": 117, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap836478e8-92", "ovs_interfaceid": "836478e8-9299-44c3-a4b0-957cd1706387", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 946.053138] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c0bbb971-43e0-418c-a0f2-fe7bf1c6b826 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.067150] env[63202]: DEBUG oslo_concurrency.lockutils [req-b92b834c-45c1-43a5-9bf0-65789a3827b5 req-3c4c7877-569f-47c5-96a4-4933f2d51a54 service nova] Acquired lock "refresh_cache-a961b2a0-39c1-4267-9229-068e2b6ecd67" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.067389] env[63202]: DEBUG nova.network.neutron [req-b92b834c-45c1-43a5-9bf0-65789a3827b5 req-3c4c7877-569f-47c5-96a4-4933f2d51a54 service nova] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Refreshing network info cache for port 836478e8-9299-44c3-a4b0-957cd1706387 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 946.068771] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b0:50:ca', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dced2f3d-7fd3-4a42-836d-9f02dab4c949', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '836478e8-9299-44c3-a4b0-957cd1706387', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 946.076106] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Creating folder: Project (f651d18a9559499da27f301ee75bb2fb). Parent ref: group-v294090. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 946.077191] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7f5b1b86-c45d-4d4f-9b72-4019962eecc7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.084121] env[63202]: DEBUG oslo_vmware.api [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for the task: (returnval){ [ 946.084121] env[63202]: value = "task-1385579" [ 946.084121] env[63202]: _type = "Task" [ 946.084121] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.088963] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Created folder: Project (f651d18a9559499da27f301ee75bb2fb) in parent group-v294090. [ 946.089214] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Creating folder: Instances. Parent ref: group-v294169. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 946.089497] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8ace9106-a78e-435a-bd9c-8959cfdcae47 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.094372] env[63202]: DEBUG oslo_vmware.api [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385579, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.101824] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Created folder: Instances in parent group-v294169. [ 946.102205] env[63202]: DEBUG oslo.service.loopingcall [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 946.102444] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 946.102787] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d750ee2a-47c2-41b2-a86d-acbb96e8770b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.130593] env[63202]: DEBUG oslo_vmware.api [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385575, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.134190] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 946.134190] env[63202]: value = "task-1385582" [ 946.134190] env[63202]: _type = "Task" [ 946.134190] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.134498] env[63202]: DEBUG oslo_vmware.api [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Task: {'id': task-1385576, 'name': Rename_Task, 'duration_secs': 0.147335} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.134749] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 946.137569] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9a977d35-924b-41ab-a168-2bbf8f834c10 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.143614] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385582, 'name': CreateVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.144738] env[63202]: DEBUG oslo_vmware.api [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Waiting for the task: (returnval){ [ 946.144738] env[63202]: value = "task-1385583" [ 946.144738] env[63202]: _type = "Task" [ 946.144738] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.154362] env[63202]: DEBUG oslo_vmware.api [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Task: {'id': task-1385583, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.159242] env[63202]: DEBUG oslo_vmware.api [None req-7de759cc-c39d-4058-8625-09664c11b490 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385577, 'name': PowerOffVM_Task, 'duration_secs': 0.404275} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.159464] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-7de759cc-c39d-4058-8625-09664c11b490 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 946.159630] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-7de759cc-c39d-4058-8625-09664c11b490 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 946.159889] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d377bdb4-72a3-47db-a7a0-78f521ad8c68 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.255758] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-7de759cc-c39d-4058-8625-09664c11b490 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 946.256076] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-7de759cc-c39d-4058-8625-09664c11b490 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Deleting contents of the VM from datastore datastore2 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 946.256313] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-7de759cc-c39d-4058-8625-09664c11b490 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Deleting the datastore file [datastore2] 3d1518d3-bed2-4b7b-af1d-d7da49e92874 {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 946.256639] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b9a80235-c3b1-4f04-936b-2acc3eb5e39c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.262978] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4e400d69-10f2-4327-9888-f1914bbefb0f tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "8755bceb-d510-4429-bd98-d6a63faf739d" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.263401] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4e400d69-10f2-4327-9888-f1914bbefb0f tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "8755bceb-d510-4429-bd98-d6a63faf739d" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.263718] env[63202]: INFO nova.compute.manager [None req-4e400d69-10f2-4327-9888-f1914bbefb0f tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Attaching volume b7ce888f-0703-4bd3-89f9-198e09e4dc43 to /dev/sdb [ 946.265606] env[63202]: DEBUG oslo_vmware.api [None req-7de759cc-c39d-4058-8625-09664c11b490 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Waiting for the task: (returnval){ [ 946.265606] env[63202]: value = "task-1385585" [ 946.265606] env[63202]: _type = "Task" [ 946.265606] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.276592] env[63202]: DEBUG oslo_vmware.api [None req-7de759cc-c39d-4058-8625-09664c11b490 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385585, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.299398] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acf77f38-f8d7-4ebd-819e-402541a66ce2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.305870] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a1f5188-6afa-419a-ba7c-2441f85f08ee {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.318839] env[63202]: DEBUG nova.virt.block_device [None req-4e400d69-10f2-4327-9888-f1914bbefb0f tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Updating existing volume attachment record: faafb4c9-1cef-4a7b-9498-20859a138879 {{(pid=63202) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 946.355496] env[63202]: DEBUG nova.scheduler.client.report [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 946.375162] env[63202]: DEBUG nova.network.neutron [-] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.543690] env[63202]: DEBUG oslo_vmware.api [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': task-1385578, 'name': Rename_Task, 'duration_secs': 0.16388} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.543972] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 946.544265] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-58fac558-5b0d-4bc8-b232-ab8168cf2b24 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.550379] env[63202]: DEBUG oslo_vmware.api [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Waiting for the task: (returnval){ [ 946.550379] env[63202]: value = "task-1385589" [ 946.550379] env[63202]: _type = "Task" [ 946.550379] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.559993] env[63202]: DEBUG oslo_vmware.api [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': task-1385589, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.595939] env[63202]: DEBUG oslo_vmware.api [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385579, 'name': ReconfigVM_Task, 'duration_secs': 0.175276} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.596316] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Reconfigured VM instance instance-00000048 to detach disk 2000 {{(pid=63202) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 946.597348] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7398c416-b46c-42b1-abe2-36faa6d394f6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.621519] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Reconfiguring VM instance instance-00000048 to attach disk [datastore1] e775e5e1-521a-4fc7-80e6-bcb6a70516c5/e775e5e1-521a-4fc7-80e6-bcb6a70516c5.vmdk or device None with type thin {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 946.624643] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-073d1987-ab03-4374-bced-6d14c38a944a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.658886] env[63202]: DEBUG oslo_vmware.api [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385575, 'name': ReconfigVM_Task, 'duration_secs': 0.538659} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.660132] env[63202]: DEBUG oslo_vmware.api [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for the task: (returnval){ [ 946.660132] env[63202]: value = "task-1385590" [ 946.660132] env[63202]: _type = "Task" [ 946.660132] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.660132] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385582, 'name': CreateVM_Task, 'duration_secs': 0.347335} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.660132] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Reconfigured VM instance instance-00000051 to attach disk [datastore2] 7437595c-fa35-483e-95f3-b75405b6bd13/7437595c-fa35-483e-95f3-b75405b6bd13.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 946.660703] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 946.660940] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6172ea76-18f0-419c-985b-fdda2ad1d9a4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.666215] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.667752] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.667752] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 946.671341] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-793bb5f0-0e5d-4244-babc-69a21aaffa47 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.673459] env[63202]: DEBUG oslo_vmware.api [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Task: {'id': task-1385583, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.677580] env[63202]: DEBUG oslo_vmware.api [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Waiting for the task: (returnval){ [ 946.677580] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]5216ec73-0cb2-961b-9db4-e6eb70895696" [ 946.677580] env[63202]: _type = "Task" [ 946.677580] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.681301] env[63202]: DEBUG oslo_vmware.api [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385590, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.685681] env[63202]: DEBUG oslo_vmware.api [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 946.685681] env[63202]: value = "task-1385591" [ 946.685681] env[63202]: _type = "Task" [ 946.685681] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.693052] env[63202]: DEBUG oslo_vmware.api [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]5216ec73-0cb2-961b-9db4-e6eb70895696, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.701581] env[63202]: DEBUG oslo_vmware.api [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385591, 'name': Rename_Task} progress is 10%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.776661] env[63202]: DEBUG oslo_vmware.api [None req-7de759cc-c39d-4058-8625-09664c11b490 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385585, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.17988} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.776935] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-7de759cc-c39d-4058-8625-09664c11b490 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 946.777139] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-7de759cc-c39d-4058-8625-09664c11b490 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Deleted contents of the VM from datastore datastore2 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 946.777326] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-7de759cc-c39d-4058-8625-09664c11b490 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 946.777495] env[63202]: INFO nova.compute.manager [None req-7de759cc-c39d-4058-8625-09664c11b490 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Took 1.15 seconds to destroy the instance on the hypervisor. [ 946.777737] env[63202]: DEBUG oslo.service.loopingcall [None req-7de759cc-c39d-4058-8625-09664c11b490 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 946.777928] env[63202]: DEBUG nova.compute.manager [-] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 946.778049] env[63202]: DEBUG nova.network.neutron [-] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 946.858679] env[63202]: DEBUG nova.network.neutron [req-b92b834c-45c1-43a5-9bf0-65789a3827b5 req-3c4c7877-569f-47c5-96a4-4933f2d51a54 service nova] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Updated VIF entry in instance network info cache for port 836478e8-9299-44c3-a4b0-957cd1706387. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 946.859092] env[63202]: DEBUG nova.network.neutron [req-b92b834c-45c1-43a5-9bf0-65789a3827b5 req-3c4c7877-569f-47c5-96a4-4933f2d51a54 service nova] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Updating instance_info_cache with network_info: [{"id": "836478e8-9299-44c3-a4b0-957cd1706387", "address": "fa:16:3e:b0:50:ca", "network": {"id": "69295efd-3c50-45e5-bc09-4f0cef6b8f0e", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1062402518-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f651d18a9559499da27f301ee75bb2fb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dced2f3d-7fd3-4a42-836d-9f02dab4c949", "external-id": "nsx-vlan-transportzone-117", "segmentation_id": 117, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap836478e8-92", "ovs_interfaceid": "836478e8-9299-44c3-a4b0-957cd1706387", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.860940] env[63202]: DEBUG oslo_concurrency.lockutils [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.452s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.861539] env[63202]: DEBUG nova.compute.manager [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 946.864697] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3c8a2d3a-2d9f-4e04-858f-39e507432c13 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.052s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.864983] env[63202]: DEBUG nova.objects.instance [None req-3c8a2d3a-2d9f-4e04-858f-39e507432c13 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Lazy-loading 'resources' on Instance uuid 753d190b-f4a4-4438-bc98-94564e3ec73d {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 946.877641] env[63202]: INFO nova.compute.manager [-] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Took 1.27 seconds to deallocate network for instance. [ 947.060131] env[63202]: DEBUG oslo_vmware.api [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': task-1385589, 'name': PowerOnVM_Task, 'duration_secs': 0.480789} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.060477] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 947.060646] env[63202]: DEBUG nova.compute.manager [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 947.061464] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0cf4ca3-0ed1-4365-92e0-d7fc11d17b3d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.154870] env[63202]: DEBUG oslo_vmware.api [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Task: {'id': task-1385583, 'name': PowerOnVM_Task, 'duration_secs': 0.539982} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.159915] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 947.159915] env[63202]: INFO nova.compute.manager [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Took 11.52 seconds to spawn the instance on the hypervisor. [ 947.159915] env[63202]: DEBUG nova.compute.manager [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 947.159915] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f239d28-e9d9-478b-81c7-88be9aaa13c6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.174834] env[63202]: DEBUG oslo_vmware.api [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385590, 'name': ReconfigVM_Task, 'duration_secs': 0.292491} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.175245] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Reconfigured VM instance instance-00000048 to attach disk [datastore1] e775e5e1-521a-4fc7-80e6-bcb6a70516c5/e775e5e1-521a-4fc7-80e6-bcb6a70516c5.vmdk or device None with type thin {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 947.175692] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Updating instance 'e775e5e1-521a-4fc7-80e6-bcb6a70516c5' progress to 50 {{(pid=63202) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 947.193640] env[63202]: DEBUG oslo_vmware.api [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]5216ec73-0cb2-961b-9db4-e6eb70895696, 'name': SearchDatastore_Task, 'duration_secs': 0.020475} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.195257] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.195486] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 947.195718] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.195855] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.196040] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 947.197207] env[63202]: DEBUG nova.compute.manager [req-c3246e97-37f1-4241-846f-4827922fa8a5 req-f2177afc-f84d-46ba-b738-54de76fcff07 service nova] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Received event network-vif-deleted-06e55a71-f982-49f8-819f-3408e19025de {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 947.197398] env[63202]: INFO nova.compute.manager [req-c3246e97-37f1-4241-846f-4827922fa8a5 req-f2177afc-f84d-46ba-b738-54de76fcff07 service nova] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Neutron deleted interface 06e55a71-f982-49f8-819f-3408e19025de; detaching it from the instance and deleting it from the info cache [ 947.197570] env[63202]: DEBUG nova.network.neutron [req-c3246e97-37f1-4241-846f-4827922fa8a5 req-f2177afc-f84d-46ba-b738-54de76fcff07 service nova] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.201928] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-40964378-c036-4b11-8ea4-0a7cee927e22 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.203843] env[63202]: DEBUG oslo_vmware.api [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385591, 'name': Rename_Task, 'duration_secs': 0.191738} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.205017] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 947.205594] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dbd93ba6-10cb-4f7f-a053-b9e25a52f08c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.211799] env[63202]: DEBUG oslo_vmware.api [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 947.211799] env[63202]: value = "task-1385592" [ 947.211799] env[63202]: _type = "Task" [ 947.211799] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.212875] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 947.213067] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 947.216853] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-21e2f592-60fc-467e-9f4b-fb2073dd2454 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.222578] env[63202]: DEBUG oslo_vmware.api [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Waiting for the task: (returnval){ [ 947.222578] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52e2e3c5-4c02-5184-4662-d9565c89edc3" [ 947.222578] env[63202]: _type = "Task" [ 947.222578] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.227670] env[63202]: DEBUG oslo_vmware.api [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385592, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.237725] env[63202]: DEBUG oslo_vmware.api [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52e2e3c5-4c02-5184-4662-d9565c89edc3, 'name': SearchDatastore_Task, 'duration_secs': 0.011147} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.238559] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0a57e67a-4976-4173-a316-aa96ef8cce5c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.244394] env[63202]: DEBUG oslo_vmware.api [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Waiting for the task: (returnval){ [ 947.244394] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52066a81-f901-56dd-e530-90f2a766b1aa" [ 947.244394] env[63202]: _type = "Task" [ 947.244394] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.253073] env[63202]: DEBUG oslo_vmware.api [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52066a81-f901-56dd-e530-90f2a766b1aa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.365746] env[63202]: DEBUG oslo_concurrency.lockutils [req-b92b834c-45c1-43a5-9bf0-65789a3827b5 req-3c4c7877-569f-47c5-96a4-4933f2d51a54 service nova] Releasing lock "refresh_cache-a961b2a0-39c1-4267-9229-068e2b6ecd67" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.367177] env[63202]: DEBUG nova.compute.utils [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 947.373015] env[63202]: DEBUG nova.compute.manager [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 947.373198] env[63202]: DEBUG nova.network.neutron [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 947.385851] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d316d71b-590a-440a-b0fc-2e046c2315ac tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.436743] env[63202]: DEBUG nova.policy [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2bd91973e1bb4ce7be8bce9bfe34a941', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e972faeaedc6468aab7e7cfee88a477b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 947.586084] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.658217] env[63202]: DEBUG nova.network.neutron [-] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.663200] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49b23e70-7cd6-46f5-a74d-526d1e87c878 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.676588] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b476dc17-4667-4ed4-b4c9-750fcef10871 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.682133] env[63202]: INFO nova.compute.manager [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Took 32.02 seconds to build instance. [ 947.685863] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a425a3d3-78e6-4f22-9ec3-021cd67d1386 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.719247] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b87d3b06-57b0-44a9-8e80-1b1ea946bcb5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.724144] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-304ba660-9f64-455e-aa96-7b59e8ad742d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.751815] env[63202]: DEBUG nova.compute.manager [req-36ca8d4f-03a8-4bf1-bab4-604acf82ccf0 req-bdd2e0f2-eb9f-487b-8c76-368f810bdf39 service nova] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Received event network-vif-deleted-41ce38c0-e94e-4de9-9504-0aaf673e3d03 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 947.751815] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb909e90-3196-4fc2-9a1e-0becfee12f37 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.764519] env[63202]: DEBUG oslo_vmware.api [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385592, 'name': PowerOnVM_Task} progress is 87%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.768385] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8d8fee1-685f-4d74-93d5-31793ff781fc {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.779903] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0af1c810-6668-4c0e-98ef-d5e08b2420d5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.798643] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Updating instance 'e775e5e1-521a-4fc7-80e6-bcb6a70516c5' progress to 67 {{(pid=63202) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 947.807481] env[63202]: DEBUG oslo_vmware.api [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52066a81-f901-56dd-e530-90f2a766b1aa, 'name': SearchDatastore_Task, 'duration_secs': 0.010901} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.807801] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.808151] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] a961b2a0-39c1-4267-9229-068e2b6ecd67/a961b2a0-39c1-4267-9229-068e2b6ecd67.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 947.815757] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-89c5f19f-e152-4b08-b7b7-755aefd25c1f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.818402] env[63202]: DEBUG nova.compute.provider_tree [None req-3c8a2d3a-2d9f-4e04-858f-39e507432c13 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 947.831527] env[63202]: DEBUG nova.compute.manager [req-c3246e97-37f1-4241-846f-4827922fa8a5 req-f2177afc-f84d-46ba-b738-54de76fcff07 service nova] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Detach interface failed, port_id=06e55a71-f982-49f8-819f-3408e19025de, reason: Instance 3d1518d3-bed2-4b7b-af1d-d7da49e92874 could not be found. {{(pid=63202) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 947.836530] env[63202]: DEBUG nova.scheduler.client.report [None req-3c8a2d3a-2d9f-4e04-858f-39e507432c13 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 947.840877] env[63202]: DEBUG oslo_vmware.api [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Waiting for the task: (returnval){ [ 947.840877] env[63202]: value = "task-1385593" [ 947.840877] env[63202]: _type = "Task" [ 947.840877] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.851577] env[63202]: DEBUG oslo_vmware.api [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385593, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.876090] env[63202]: DEBUG nova.compute.manager [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 947.886316] env[63202]: DEBUG nova.network.neutron [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Successfully created port: 7d59550a-39cc-4a17-a5a8-a03665c6914a {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 948.160524] env[63202]: INFO nova.compute.manager [-] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Took 1.38 seconds to deallocate network for instance. [ 948.192627] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c8dcb3c4-445e-4baa-be9b-c3883c28cabc tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Lock "f3128c09-3680-4b0b-b463-3d6cd203fcf4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.679s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.227414] env[63202]: DEBUG oslo_vmware.api [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385592, 'name': PowerOnVM_Task, 'duration_secs': 0.841985} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.227670] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 948.227873] env[63202]: INFO nova.compute.manager [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Took 9.26 seconds to spawn the instance on the hypervisor. [ 948.228065] env[63202]: DEBUG nova.compute.manager [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 948.229099] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a2e4f01-5e0f-4495-a51f-ba38d4a938a4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.248971] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f52d6525-76d4-472b-8523-a728a64d8d38 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Acquiring lock "fc7d3924-a624-4ea4-890c-7628595fb733" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.249176] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f52d6525-76d4-472b-8523-a728a64d8d38 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Lock "fc7d3924-a624-4ea4-890c-7628595fb733" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.249380] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f52d6525-76d4-472b-8523-a728a64d8d38 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Acquiring lock "fc7d3924-a624-4ea4-890c-7628595fb733-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.249857] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f52d6525-76d4-472b-8523-a728a64d8d38 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Lock "fc7d3924-a624-4ea4-890c-7628595fb733-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.250054] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f52d6525-76d4-472b-8523-a728a64d8d38 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Lock "fc7d3924-a624-4ea4-890c-7628595fb733-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.251999] env[63202]: INFO nova.compute.manager [None req-f52d6525-76d4-472b-8523-a728a64d8d38 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Terminating instance [ 948.253602] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f52d6525-76d4-472b-8523-a728a64d8d38 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Acquiring lock "refresh_cache-fc7d3924-a624-4ea4-890c-7628595fb733" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.253755] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f52d6525-76d4-472b-8523-a728a64d8d38 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Acquired lock "refresh_cache-fc7d3924-a624-4ea4-890c-7628595fb733" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.253919] env[63202]: DEBUG nova.network.neutron [None req-f52d6525-76d4-472b-8523-a728a64d8d38 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 948.338602] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3c8a2d3a-2d9f-4e04-858f-39e507432c13 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.474s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.341637] env[63202]: DEBUG nova.network.neutron [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Port c7da6929-1bd1-4e2b-bcb8-609863cf6861 binding to destination host cpu-1 is already ACTIVE {{(pid=63202) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 948.343030] env[63202]: DEBUG oslo_concurrency.lockutils [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.391s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.344731] env[63202]: INFO nova.compute.claims [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: da285417-bd38-4387-8521-df326ca0e326] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 948.358088] env[63202]: DEBUG oslo_vmware.api [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385593, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.365428] env[63202]: INFO nova.scheduler.client.report [None req-3c8a2d3a-2d9f-4e04-858f-39e507432c13 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Deleted allocations for instance 753d190b-f4a4-4438-bc98-94564e3ec73d [ 948.666899] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7de759cc-c39d-4058-8625-09664c11b490 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.746408] env[63202]: INFO nova.compute.manager [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Took 30.41 seconds to build instance. [ 948.780386] env[63202]: DEBUG nova.network.neutron [None req-f52d6525-76d4-472b-8523-a728a64d8d38 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 948.837890] env[63202]: DEBUG nova.network.neutron [None req-f52d6525-76d4-472b-8523-a728a64d8d38 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.866852] env[63202]: DEBUG oslo_vmware.api [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385593, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.82588} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.867133] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] a961b2a0-39c1-4267-9229-068e2b6ecd67/a961b2a0-39c1-4267-9229-068e2b6ecd67.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 948.867347] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 948.867596] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0d4e830c-aea9-46d8-b6bf-e8d5578c1348 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.874040] env[63202]: DEBUG oslo_vmware.api [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Waiting for the task: (returnval){ [ 948.874040] env[63202]: value = "task-1385595" [ 948.874040] env[63202]: _type = "Task" [ 948.874040] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.882415] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3c8a2d3a-2d9f-4e04-858f-39e507432c13 tempest-ServerAddressesNegativeTestJSON-1047795468 tempest-ServerAddressesNegativeTestJSON-1047795468-project-member] Lock "753d190b-f4a4-4438-bc98-94564e3ec73d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.716s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.887944] env[63202]: DEBUG oslo_vmware.api [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385595, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.891602] env[63202]: DEBUG nova.compute.manager [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 948.923845] env[63202]: DEBUG nova.virt.hardware [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 948.924112] env[63202]: DEBUG nova.virt.hardware [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 948.924317] env[63202]: DEBUG nova.virt.hardware [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 948.924502] env[63202]: DEBUG nova.virt.hardware [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 948.924659] env[63202]: DEBUG nova.virt.hardware [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 948.924812] env[63202]: DEBUG nova.virt.hardware [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 948.925059] env[63202]: DEBUG nova.virt.hardware [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 948.925232] env[63202]: DEBUG nova.virt.hardware [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 948.925420] env[63202]: DEBUG nova.virt.hardware [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 948.925602] env[63202]: DEBUG nova.virt.hardware [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 948.925835] env[63202]: DEBUG nova.virt.hardware [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 948.926943] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71036137-1e3a-4786-af0f-2b1bfcdb4886 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.938278] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a4057f2-4dfc-4c39-83dd-8e4bb6af3021 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.250163] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e6f93955-d69b-41d8-900b-9a5e95489aa2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "7437595c-fa35-483e-95f3-b75405b6bd13" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.816s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.340975] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f52d6525-76d4-472b-8523-a728a64d8d38 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Releasing lock "refresh_cache-fc7d3924-a624-4ea4-890c-7628595fb733" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.341392] env[63202]: DEBUG nova.compute.manager [None req-f52d6525-76d4-472b-8523-a728a64d8d38 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 949.341573] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-f52d6525-76d4-472b-8523-a728a64d8d38 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 949.342558] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-634136ee-10a3-4d56-a4cf-f9ac92164b67 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.351104] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-f52d6525-76d4-472b-8523-a728a64d8d38 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 949.351407] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b7d43724-48c1-41ac-9e76-e0026bd990c9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.360879] env[63202]: DEBUG oslo_vmware.api [None req-f52d6525-76d4-472b-8523-a728a64d8d38 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Waiting for the task: (returnval){ [ 949.360879] env[63202]: value = "task-1385596" [ 949.360879] env[63202]: _type = "Task" [ 949.360879] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.379979] env[63202]: DEBUG nova.compute.manager [req-87e7a318-ad3c-40de-a1da-f4b55860c80e req-e199ac69-cfef-44fb-9358-afdd36f780d2 service nova] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Received event network-changed-be85354c-2bf7-4c06-898f-7061cdfada23 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 949.379979] env[63202]: DEBUG nova.compute.manager [req-87e7a318-ad3c-40de-a1da-f4b55860c80e req-e199ac69-cfef-44fb-9358-afdd36f780d2 service nova] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Refreshing instance network info cache due to event network-changed-be85354c-2bf7-4c06-898f-7061cdfada23. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 949.379979] env[63202]: DEBUG oslo_concurrency.lockutils [req-87e7a318-ad3c-40de-a1da-f4b55860c80e req-e199ac69-cfef-44fb-9358-afdd36f780d2 service nova] Acquiring lock "refresh_cache-f3128c09-3680-4b0b-b463-3d6cd203fcf4" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 949.379979] env[63202]: DEBUG oslo_concurrency.lockutils [req-87e7a318-ad3c-40de-a1da-f4b55860c80e req-e199ac69-cfef-44fb-9358-afdd36f780d2 service nova] Acquired lock "refresh_cache-f3128c09-3680-4b0b-b463-3d6cd203fcf4" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.379979] env[63202]: DEBUG nova.network.neutron [req-87e7a318-ad3c-40de-a1da-f4b55860c80e req-e199ac69-cfef-44fb-9358-afdd36f780d2 service nova] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Refreshing network info cache for port be85354c-2bf7-4c06-898f-7061cdfada23 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 949.394169] env[63202]: DEBUG oslo_vmware.api [None req-f52d6525-76d4-472b-8523-a728a64d8d38 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': task-1385596, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.397197] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "e775e5e1-521a-4fc7-80e6-bcb6a70516c5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.397426] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "e775e5e1-521a-4fc7-80e6-bcb6a70516c5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.397592] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "e775e5e1-521a-4fc7-80e6-bcb6a70516c5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.405272] env[63202]: DEBUG oslo_vmware.api [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385595, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063591} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.405832] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 949.406362] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1ffe18f-e832-47ee-954f-fd8007bea8a8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.430302] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Reconfiguring VM instance instance-00000052 to attach disk [datastore1] a961b2a0-39c1-4267-9229-068e2b6ecd67/a961b2a0-39c1-4267-9229-068e2b6ecd67.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 949.433849] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-833d85f8-c765-4993-abc5-8d93eafce8d0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.457017] env[63202]: DEBUG oslo_vmware.api [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Waiting for the task: (returnval){ [ 949.457017] env[63202]: value = "task-1385597" [ 949.457017] env[63202]: _type = "Task" [ 949.457017] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.468515] env[63202]: DEBUG oslo_vmware.api [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385597, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.695084] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca23972c-4d17-4c28-bf92-77ac49874f19 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.704247] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc391922-4c0d-426e-aa28-9c74bfbfab75 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.738831] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73434d79-ec9b-44f1-a535-3a4375b76313 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.748936] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd541350-ca4d-4a2b-8d54-7deb1377a3a2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.764905] env[63202]: DEBUG nova.compute.provider_tree [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Updating inventory in ProviderTree for provider 79b33d17-3e75-494c-a550-67b275de2079 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 949.843635] env[63202]: DEBUG nova.compute.manager [req-65fb32b9-f08a-44ce-9aa1-d3ced39df12a req-571562f3-b79a-4281-b307-cf390ec191d2 service nova] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Received event network-changed-5eeb4ae8-9356-402a-965d-224788f04129 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 949.843770] env[63202]: DEBUG nova.compute.manager [req-65fb32b9-f08a-44ce-9aa1-d3ced39df12a req-571562f3-b79a-4281-b307-cf390ec191d2 service nova] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Refreshing instance network info cache due to event network-changed-5eeb4ae8-9356-402a-965d-224788f04129. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 949.843983] env[63202]: DEBUG oslo_concurrency.lockutils [req-65fb32b9-f08a-44ce-9aa1-d3ced39df12a req-571562f3-b79a-4281-b307-cf390ec191d2 service nova] Acquiring lock "refresh_cache-7437595c-fa35-483e-95f3-b75405b6bd13" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 949.844133] env[63202]: DEBUG oslo_concurrency.lockutils [req-65fb32b9-f08a-44ce-9aa1-d3ced39df12a req-571562f3-b79a-4281-b307-cf390ec191d2 service nova] Acquired lock "refresh_cache-7437595c-fa35-483e-95f3-b75405b6bd13" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.844292] env[63202]: DEBUG nova.network.neutron [req-65fb32b9-f08a-44ce-9aa1-d3ced39df12a req-571562f3-b79a-4281-b307-cf390ec191d2 service nova] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Refreshing network info cache for port 5eeb4ae8-9356-402a-965d-224788f04129 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 949.880323] env[63202]: DEBUG oslo_vmware.api [None req-f52d6525-76d4-472b-8523-a728a64d8d38 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': task-1385596, 'name': PowerOffVM_Task, 'duration_secs': 0.125972} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.880323] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-f52d6525-76d4-472b-8523-a728a64d8d38 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 949.880323] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-f52d6525-76d4-472b-8523-a728a64d8d38 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 949.880323] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e8330aaf-61dd-4c3b-ad8c-55feb3695c27 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.914218] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-f52d6525-76d4-472b-8523-a728a64d8d38 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 949.914471] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-f52d6525-76d4-472b-8523-a728a64d8d38 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Deleting contents of the VM from datastore datastore2 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 949.914662] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-f52d6525-76d4-472b-8523-a728a64d8d38 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Deleting the datastore file [datastore2] fc7d3924-a624-4ea4-890c-7628595fb733 {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 949.916274] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bd9f9a97-5726-45a3-beeb-7f2d9de3391d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.924580] env[63202]: DEBUG oslo_vmware.api [None req-f52d6525-76d4-472b-8523-a728a64d8d38 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Waiting for the task: (returnval){ [ 949.924580] env[63202]: value = "task-1385599" [ 949.924580] env[63202]: _type = "Task" [ 949.924580] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.940363] env[63202]: DEBUG oslo_vmware.api [None req-f52d6525-76d4-472b-8523-a728a64d8d38 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': task-1385599, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.969075] env[63202]: DEBUG oslo_vmware.api [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385597, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.974190] env[63202]: DEBUG nova.network.neutron [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Successfully updated port: 7d59550a-39cc-4a17-a5a8-a03665c6914a {{(pid=63202) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 950.272970] env[63202]: DEBUG nova.network.neutron [req-87e7a318-ad3c-40de-a1da-f4b55860c80e req-e199ac69-cfef-44fb-9358-afdd36f780d2 service nova] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Updated VIF entry in instance network info cache for port be85354c-2bf7-4c06-898f-7061cdfada23. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 950.273381] env[63202]: DEBUG nova.network.neutron [req-87e7a318-ad3c-40de-a1da-f4b55860c80e req-e199ac69-cfef-44fb-9358-afdd36f780d2 service nova] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Updating instance_info_cache with network_info: [{"id": "be85354c-2bf7-4c06-898f-7061cdfada23", "address": "fa:16:3e:e6:8e:d3", "network": {"id": "8a3e2159-c8ca-4ad6-b9c6-87029c42fa61", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-503250460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.221", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "417f815aec0c451a8ccd713ad6c64ad2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbe85354c-2b", "ovs_interfaceid": "be85354c-2bf7-4c06-898f-7061cdfada23", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.302291] env[63202]: DEBUG nova.scheduler.client.report [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Updated inventory for provider 79b33d17-3e75-494c-a550-67b275de2079 with generation 95 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 950.302557] env[63202]: DEBUG nova.compute.provider_tree [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Updating resource provider 79b33d17-3e75-494c-a550-67b275de2079 generation from 95 to 96 during operation: update_inventory {{(pid=63202) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 950.303149] env[63202]: DEBUG nova.compute.provider_tree [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Updating inventory in ProviderTree for provider 79b33d17-3e75-494c-a550-67b275de2079 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 950.437499] env[63202]: DEBUG oslo_vmware.api [None req-f52d6525-76d4-472b-8523-a728a64d8d38 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Task: {'id': task-1385599, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.237564} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.437759] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-f52d6525-76d4-472b-8523-a728a64d8d38 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 950.437944] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-f52d6525-76d4-472b-8523-a728a64d8d38 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Deleted contents of the VM from datastore datastore2 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 950.438477] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-f52d6525-76d4-472b-8523-a728a64d8d38 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 950.438683] env[63202]: INFO nova.compute.manager [None req-f52d6525-76d4-472b-8523-a728a64d8d38 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Took 1.10 seconds to destroy the instance on the hypervisor. [ 950.438935] env[63202]: DEBUG oslo.service.loopingcall [None req-f52d6525-76d4-472b-8523-a728a64d8d38 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 950.439157] env[63202]: DEBUG nova.compute.manager [-] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 950.439250] env[63202]: DEBUG nova.network.neutron [-] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 950.454330] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "refresh_cache-e775e5e1-521a-4fc7-80e6-bcb6a70516c5" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.454514] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquired lock "refresh_cache-e775e5e1-521a-4fc7-80e6-bcb6a70516c5" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.454699] env[63202]: DEBUG nova.network.neutron [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 950.473373] env[63202]: DEBUG oslo_vmware.api [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385597, 'name': ReconfigVM_Task, 'duration_secs': 0.608489} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.473990] env[63202]: DEBUG nova.network.neutron [-] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 950.476366] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Reconfigured VM instance instance-00000052 to attach disk [datastore1] a961b2a0-39c1-4267-9229-068e2b6ecd67/a961b2a0-39c1-4267-9229-068e2b6ecd67.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 950.477426] env[63202]: DEBUG oslo_concurrency.lockutils [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "refresh_cache-2110dfa7-1795-4c8f-92bf-18921409c99f" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.477563] env[63202]: DEBUG oslo_concurrency.lockutils [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquired lock "refresh_cache-2110dfa7-1795-4c8f-92bf-18921409c99f" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.477700] env[63202]: DEBUG nova.network.neutron [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 950.478613] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-000c8c56-2602-45c0-a26a-cc0275497b13 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.488343] env[63202]: DEBUG oslo_vmware.api [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Waiting for the task: (returnval){ [ 950.488343] env[63202]: value = "task-1385600" [ 950.488343] env[63202]: _type = "Task" [ 950.488343] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.503897] env[63202]: DEBUG oslo_vmware.api [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385600, 'name': Rename_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.660531] env[63202]: DEBUG nova.network.neutron [req-65fb32b9-f08a-44ce-9aa1-d3ced39df12a req-571562f3-b79a-4281-b307-cf390ec191d2 service nova] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Updated VIF entry in instance network info cache for port 5eeb4ae8-9356-402a-965d-224788f04129. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 950.661230] env[63202]: DEBUG nova.network.neutron [req-65fb32b9-f08a-44ce-9aa1-d3ced39df12a req-571562f3-b79a-4281-b307-cf390ec191d2 service nova] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Updating instance_info_cache with network_info: [{"id": "5eeb4ae8-9356-402a-965d-224788f04129", "address": "fa:16:3e:b1:37:8e", "network": {"id": "0aa55fee-953e-4c2d-b88a-43b0cf86c7f3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-878071414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.153", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7d28fea097541adadf0839940568409", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5eeb4ae8-93", "ovs_interfaceid": "5eeb4ae8-9356-402a-965d-224788f04129", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.779129] env[63202]: DEBUG oslo_concurrency.lockutils [req-87e7a318-ad3c-40de-a1da-f4b55860c80e req-e199ac69-cfef-44fb-9358-afdd36f780d2 service nova] Releasing lock "refresh_cache-f3128c09-3680-4b0b-b463-3d6cd203fcf4" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.807558] env[63202]: DEBUG oslo_concurrency.lockutils [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.464s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.808198] env[63202]: DEBUG nova.compute.manager [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: da285417-bd38-4387-8521-df326ca0e326] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 950.813027] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4eda49d7-5b02-4d51-99e9-f29b250eca15 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.824s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.813027] env[63202]: DEBUG nova.objects.instance [None req-4eda49d7-5b02-4d51-99e9-f29b250eca15 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Lazy-loading 'resources' on Instance uuid af2bedc2-28ee-4679-ae38-1cceb2af05d6 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 950.871602] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e400d69-10f2-4327-9888-f1914bbefb0f tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Volume attach. Driver type: vmdk {{(pid=63202) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 950.871849] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e400d69-10f2-4327-9888-f1914bbefb0f tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294173', 'volume_id': 'b7ce888f-0703-4bd3-89f9-198e09e4dc43', 'name': 'volume-b7ce888f-0703-4bd3-89f9-198e09e4dc43', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8755bceb-d510-4429-bd98-d6a63faf739d', 'attached_at': '', 'detached_at': '', 'volume_id': 'b7ce888f-0703-4bd3-89f9-198e09e4dc43', 'serial': 'b7ce888f-0703-4bd3-89f9-198e09e4dc43'} {{(pid=63202) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 950.875541] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9b25ccf-234d-4a4c-9297-f24f95acc2df {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.897981] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce06b4e4-3a8e-450a-854a-a5db0e7f438a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.924210] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e400d69-10f2-4327-9888-f1914bbefb0f tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Reconfiguring VM instance instance-0000004f to attach disk [datastore1] volume-b7ce888f-0703-4bd3-89f9-198e09e4dc43/volume-b7ce888f-0703-4bd3-89f9-198e09e4dc43.vmdk or device None with type thin {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 950.924556] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8b6b92bd-e64d-4626-b279-5fd4d984b76f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.944254] env[63202]: DEBUG oslo_vmware.api [None req-4e400d69-10f2-4327-9888-f1914bbefb0f tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 950.944254] env[63202]: value = "task-1385601" [ 950.944254] env[63202]: _type = "Task" [ 950.944254] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.958946] env[63202]: DEBUG oslo_vmware.api [None req-4e400d69-10f2-4327-9888-f1914bbefb0f tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385601, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.976426] env[63202]: DEBUG nova.network.neutron [-] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.000476] env[63202]: DEBUG oslo_vmware.api [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385600, 'name': Rename_Task, 'duration_secs': 0.253184} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.001118] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 951.002184] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4e036dc3-9904-4101-b7fd-93a9d07ce757 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.011387] env[63202]: DEBUG oslo_vmware.api [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Waiting for the task: (returnval){ [ 951.011387] env[63202]: value = "task-1385602" [ 951.011387] env[63202]: _type = "Task" [ 951.011387] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.024769] env[63202]: DEBUG oslo_vmware.api [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385602, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.033196] env[63202]: DEBUG nova.network.neutron [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 951.167103] env[63202]: DEBUG oslo_concurrency.lockutils [req-65fb32b9-f08a-44ce-9aa1-d3ced39df12a req-571562f3-b79a-4281-b307-cf390ec191d2 service nova] Releasing lock "refresh_cache-7437595c-fa35-483e-95f3-b75405b6bd13" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.309440] env[63202]: DEBUG nova.network.neutron [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Updating instance_info_cache with network_info: [{"id": "7d59550a-39cc-4a17-a5a8-a03665c6914a", "address": "fa:16:3e:da:06:53", "network": {"id": "b1ea2320-3801-43fe-b566-4f69db81329e", "bridge": "br-int", "label": "tempest-ServersTestJSON-618582355-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e972faeaedc6468aab7e7cfee88a477b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4223acd2-30f7-440e-b975-60b30d931694", "external-id": "nsx-vlan-transportzone-647", "segmentation_id": 647, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d59550a-39", "ovs_interfaceid": "7d59550a-39cc-4a17-a5a8-a03665c6914a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.315043] env[63202]: DEBUG nova.compute.utils [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 951.319699] env[63202]: DEBUG nova.compute.manager [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: da285417-bd38-4387-8521-df326ca0e326] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 951.320660] env[63202]: DEBUG nova.network.neutron [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: da285417-bd38-4387-8521-df326ca0e326] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 951.386519] env[63202]: DEBUG nova.network.neutron [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Updating instance_info_cache with network_info: [{"id": "c7da6929-1bd1-4e2b-bcb8-609863cf6861", "address": "fa:16:3e:cf:6c:83", "network": {"id": "18390479-16ce-4012-9ba6-abf19b5616d1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1376026898-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22352c25bca8416a948014391a5389ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "085fb0ff-9285-4f1d-a008-a14da4844357", "external-id": "nsx-vlan-transportzone-729", "segmentation_id": 729, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7da6929-1b", "ovs_interfaceid": "c7da6929-1bd1-4e2b-bcb8-609863cf6861", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.392593] env[63202]: DEBUG nova.policy [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2d292997e4c34051acc645b4a65dc369', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5e914f87450949b1a39866e8cfa3a0eb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 951.408765] env[63202]: DEBUG nova.compute.manager [req-c9e0b5a3-f18d-4aa8-b682-a9eda0f49f5a req-8896d125-0a60-441b-8c71-b928070cc396 service nova] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Received event network-vif-plugged-7d59550a-39cc-4a17-a5a8-a03665c6914a {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 951.408994] env[63202]: DEBUG oslo_concurrency.lockutils [req-c9e0b5a3-f18d-4aa8-b682-a9eda0f49f5a req-8896d125-0a60-441b-8c71-b928070cc396 service nova] Acquiring lock "2110dfa7-1795-4c8f-92bf-18921409c99f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.409243] env[63202]: DEBUG oslo_concurrency.lockutils [req-c9e0b5a3-f18d-4aa8-b682-a9eda0f49f5a req-8896d125-0a60-441b-8c71-b928070cc396 service nova] Lock "2110dfa7-1795-4c8f-92bf-18921409c99f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.409432] env[63202]: DEBUG oslo_concurrency.lockutils [req-c9e0b5a3-f18d-4aa8-b682-a9eda0f49f5a req-8896d125-0a60-441b-8c71-b928070cc396 service nova] Lock "2110dfa7-1795-4c8f-92bf-18921409c99f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.409597] env[63202]: DEBUG nova.compute.manager [req-c9e0b5a3-f18d-4aa8-b682-a9eda0f49f5a req-8896d125-0a60-441b-8c71-b928070cc396 service nova] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] No waiting events found dispatching network-vif-plugged-7d59550a-39cc-4a17-a5a8-a03665c6914a {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 951.409776] env[63202]: WARNING nova.compute.manager [req-c9e0b5a3-f18d-4aa8-b682-a9eda0f49f5a req-8896d125-0a60-441b-8c71-b928070cc396 service nova] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Received unexpected event network-vif-plugged-7d59550a-39cc-4a17-a5a8-a03665c6914a for instance with vm_state building and task_state spawning. [ 951.409950] env[63202]: DEBUG nova.compute.manager [req-c9e0b5a3-f18d-4aa8-b682-a9eda0f49f5a req-8896d125-0a60-441b-8c71-b928070cc396 service nova] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Received event network-changed-7d59550a-39cc-4a17-a5a8-a03665c6914a {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 951.410118] env[63202]: DEBUG nova.compute.manager [req-c9e0b5a3-f18d-4aa8-b682-a9eda0f49f5a req-8896d125-0a60-441b-8c71-b928070cc396 service nova] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Refreshing instance network info cache due to event network-changed-7d59550a-39cc-4a17-a5a8-a03665c6914a. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 951.410317] env[63202]: DEBUG oslo_concurrency.lockutils [req-c9e0b5a3-f18d-4aa8-b682-a9eda0f49f5a req-8896d125-0a60-441b-8c71-b928070cc396 service nova] Acquiring lock "refresh_cache-2110dfa7-1795-4c8f-92bf-18921409c99f" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.456643] env[63202]: DEBUG oslo_vmware.api [None req-4e400d69-10f2-4327-9888-f1914bbefb0f tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385601, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.479267] env[63202]: INFO nova.compute.manager [-] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Took 1.04 seconds to deallocate network for instance. [ 951.526410] env[63202]: DEBUG oslo_vmware.api [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385602, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.627264] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e68cb00-c3a5-478c-b8dd-d03278e2107f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.638261] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb3fed5b-fa19-4622-b2bb-7327c3315a25 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.679898] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b185290e-5184-40a7-b522-064cc530724a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.680359] env[63202]: DEBUG nova.network.neutron [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: da285417-bd38-4387-8521-df326ca0e326] Successfully created port: a0547ee8-b263-49d2-8152-cbac29b8c25b {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 951.688772] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-767be470-78ab-48e3-9a4d-78d0f4803d00 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.705320] env[63202]: DEBUG nova.compute.provider_tree [None req-4eda49d7-5b02-4d51-99e9-f29b250eca15 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 951.818786] env[63202]: DEBUG nova.compute.manager [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: da285417-bd38-4387-8521-df326ca0e326] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 951.826323] env[63202]: DEBUG oslo_concurrency.lockutils [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Releasing lock "refresh_cache-2110dfa7-1795-4c8f-92bf-18921409c99f" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.826323] env[63202]: DEBUG nova.compute.manager [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Instance network_info: |[{"id": "7d59550a-39cc-4a17-a5a8-a03665c6914a", "address": "fa:16:3e:da:06:53", "network": {"id": "b1ea2320-3801-43fe-b566-4f69db81329e", "bridge": "br-int", "label": "tempest-ServersTestJSON-618582355-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e972faeaedc6468aab7e7cfee88a477b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4223acd2-30f7-440e-b975-60b30d931694", "external-id": "nsx-vlan-transportzone-647", "segmentation_id": 647, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d59550a-39", "ovs_interfaceid": "7d59550a-39cc-4a17-a5a8-a03665c6914a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 951.826536] env[63202]: DEBUG oslo_concurrency.lockutils [req-c9e0b5a3-f18d-4aa8-b682-a9eda0f49f5a req-8896d125-0a60-441b-8c71-b928070cc396 service nova] Acquired lock "refresh_cache-2110dfa7-1795-4c8f-92bf-18921409c99f" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.826991] env[63202]: DEBUG nova.network.neutron [req-c9e0b5a3-f18d-4aa8-b682-a9eda0f49f5a req-8896d125-0a60-441b-8c71-b928070cc396 service nova] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Refreshing network info cache for port 7d59550a-39cc-4a17-a5a8-a03665c6914a {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 951.828643] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:da:06:53', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4223acd2-30f7-440e-b975-60b30d931694', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7d59550a-39cc-4a17-a5a8-a03665c6914a', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 951.842791] env[63202]: DEBUG oslo.service.loopingcall [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 951.844113] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 951.846847] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-afcb3c90-6e57-48f2-a4e7-889e81831f9f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.879778] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 951.879778] env[63202]: value = "task-1385603" [ 951.879778] env[63202]: _type = "Task" [ 951.879778] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.891357] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385603, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.896538] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Releasing lock "refresh_cache-e775e5e1-521a-4fc7-80e6-bcb6a70516c5" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.956024] env[63202]: DEBUG oslo_vmware.api [None req-4e400d69-10f2-4327-9888-f1914bbefb0f tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385601, 'name': ReconfigVM_Task, 'duration_secs': 0.638132} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.956112] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e400d69-10f2-4327-9888-f1914bbefb0f tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Reconfigured VM instance instance-0000004f to attach disk [datastore1] volume-b7ce888f-0703-4bd3-89f9-198e09e4dc43/volume-b7ce888f-0703-4bd3-89f9-198e09e4dc43.vmdk or device None with type thin {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 951.961585] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c890012f-5d9f-43a5-8538-7cceddfd7536 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.979034] env[63202]: DEBUG oslo_vmware.api [None req-4e400d69-10f2-4327-9888-f1914bbefb0f tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 951.979034] env[63202]: value = "task-1385604" [ 951.979034] env[63202]: _type = "Task" [ 951.979034] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.988969] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f52d6525-76d4-472b-8523-a728a64d8d38 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.989646] env[63202]: DEBUG oslo_vmware.api [None req-4e400d69-10f2-4327-9888-f1914bbefb0f tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385604, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.024022] env[63202]: DEBUG oslo_vmware.api [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385602, 'name': PowerOnVM_Task, 'duration_secs': 0.789334} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.024022] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 952.024252] env[63202]: INFO nova.compute.manager [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Took 7.62 seconds to spawn the instance on the hypervisor. [ 952.024289] env[63202]: DEBUG nova.compute.manager [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 952.025113] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a4ac92c-0d86-49b5-b9f8-4a9bc07f0671 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.209152] env[63202]: DEBUG nova.scheduler.client.report [None req-4eda49d7-5b02-4d51-99e9-f29b250eca15 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 952.392068] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385603, 'name': CreateVM_Task, 'duration_secs': 0.340727} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.392363] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 952.392924] env[63202]: DEBUG oslo_concurrency.lockutils [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.393100] env[63202]: DEBUG oslo_concurrency.lockutils [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.393417] env[63202]: DEBUG oslo_concurrency.lockutils [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 952.393673] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a4f0f86d-74a6-45a2-83fa-9000cef2abd8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.398934] env[63202]: DEBUG oslo_vmware.api [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 952.398934] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52108820-d9b2-3ba8-0556-822332590fd5" [ 952.398934] env[63202]: _type = "Task" [ 952.398934] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.412929] env[63202]: DEBUG oslo_vmware.api [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52108820-d9b2-3ba8-0556-822332590fd5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.422089] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96cea32d-244f-47c3-bb94-429711af90d3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.443256] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62ee1133-08c9-46b2-a9b6-56ec67fe42f3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.450957] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Updating instance 'e775e5e1-521a-4fc7-80e6-bcb6a70516c5' progress to 83 {{(pid=63202) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 952.489616] env[63202]: DEBUG oslo_vmware.api [None req-4e400d69-10f2-4327-9888-f1914bbefb0f tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385604, 'name': ReconfigVM_Task, 'duration_secs': 0.205903} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.489927] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e400d69-10f2-4327-9888-f1914bbefb0f tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294173', 'volume_id': 'b7ce888f-0703-4bd3-89f9-198e09e4dc43', 'name': 'volume-b7ce888f-0703-4bd3-89f9-198e09e4dc43', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8755bceb-d510-4429-bd98-d6a63faf739d', 'attached_at': '', 'detached_at': '', 'volume_id': 'b7ce888f-0703-4bd3-89f9-198e09e4dc43', 'serial': 'b7ce888f-0703-4bd3-89f9-198e09e4dc43'} {{(pid=63202) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 952.544697] env[63202]: INFO nova.compute.manager [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Took 29.74 seconds to build instance. [ 952.714299] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4eda49d7-5b02-4d51-99e9-f29b250eca15 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.903s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.718708] env[63202]: DEBUG nova.network.neutron [req-c9e0b5a3-f18d-4aa8-b682-a9eda0f49f5a req-8896d125-0a60-441b-8c71-b928070cc396 service nova] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Updated VIF entry in instance network info cache for port 7d59550a-39cc-4a17-a5a8-a03665c6914a. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 952.721554] env[63202]: DEBUG nova.network.neutron [req-c9e0b5a3-f18d-4aa8-b682-a9eda0f49f5a req-8896d125-0a60-441b-8c71-b928070cc396 service nova] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Updating instance_info_cache with network_info: [{"id": "7d59550a-39cc-4a17-a5a8-a03665c6914a", "address": "fa:16:3e:da:06:53", "network": {"id": "b1ea2320-3801-43fe-b566-4f69db81329e", "bridge": "br-int", "label": "tempest-ServersTestJSON-618582355-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e972faeaedc6468aab7e7cfee88a477b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4223acd2-30f7-440e-b975-60b30d931694", "external-id": "nsx-vlan-transportzone-647", "segmentation_id": 647, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d59550a-39", "ovs_interfaceid": "7d59550a-39cc-4a17-a5a8-a03665c6914a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.723274] env[63202]: DEBUG oslo_concurrency.lockutils [None req-02dae3ed-6d57-4066-a061-72500d73ac65 tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.842s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.724425] env[63202]: DEBUG nova.objects.instance [None req-02dae3ed-6d57-4066-a061-72500d73ac65 tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Lazy-loading 'resources' on Instance uuid 2243bb69-0dc5-49cd-b94e-73e703cbadc0 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 952.742716] env[63202]: INFO nova.scheduler.client.report [None req-4eda49d7-5b02-4d51-99e9-f29b250eca15 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Deleted allocations for instance af2bedc2-28ee-4679-ae38-1cceb2af05d6 [ 952.836018] env[63202]: DEBUG nova.compute.manager [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: da285417-bd38-4387-8521-df326ca0e326] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 952.864512] env[63202]: DEBUG nova.virt.hardware [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='208106534b2f9ac5a193d33180e7fe79',container_format='bare',created_at=2024-10-31T10:28:21Z,direct_url=,disk_format='vmdk',id=376aee6c-1f98-4a60-858b-40c55f65b90e,min_disk=1,min_ram=0,name='tempest-test-snap-1118545241',owner='5e914f87450949b1a39866e8cfa3a0eb',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-10-31T10:28:38Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 952.864758] env[63202]: DEBUG nova.virt.hardware [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 952.864917] env[63202]: DEBUG nova.virt.hardware [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 952.865279] env[63202]: DEBUG nova.virt.hardware [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 952.865279] env[63202]: DEBUG nova.virt.hardware [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 952.865402] env[63202]: DEBUG nova.virt.hardware [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 952.865610] env[63202]: DEBUG nova.virt.hardware [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 952.865769] env[63202]: DEBUG nova.virt.hardware [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 952.865989] env[63202]: DEBUG nova.virt.hardware [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 952.866198] env[63202]: DEBUG nova.virt.hardware [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 952.866378] env[63202]: DEBUG nova.virt.hardware [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 952.867587] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2784190-ac1a-47eb-9a9c-89d60b414203 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.876884] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c104ce9-bc1d-44a9-8e1c-da92226af908 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.909400] env[63202]: DEBUG oslo_vmware.api [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52108820-d9b2-3ba8-0556-822332590fd5, 'name': SearchDatastore_Task, 'duration_secs': 0.018347} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.909693] env[63202]: DEBUG oslo_concurrency.lockutils [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.909947] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 952.910167] env[63202]: DEBUG oslo_concurrency.lockutils [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.910322] env[63202]: DEBUG oslo_concurrency.lockutils [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.912204] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 952.912204] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-25168df4-9796-4559-8c01-38381a7b08d3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.924035] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 952.924227] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 952.924926] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1d061151-6ed8-475a-9575-ef3d9c7ab9a7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.930172] env[63202]: DEBUG oslo_vmware.api [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 952.930172] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52775c09-20e1-e588-7c16-f6a321278364" [ 952.930172] env[63202]: _type = "Task" [ 952.930172] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.938125] env[63202]: DEBUG oslo_vmware.api [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52775c09-20e1-e588-7c16-f6a321278364, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.960157] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 952.960479] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ccc66cc0-14d1-42b7-b451-b0d84706d0f8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.970480] env[63202]: DEBUG oslo_vmware.api [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for the task: (returnval){ [ 952.970480] env[63202]: value = "task-1385605" [ 952.970480] env[63202]: _type = "Task" [ 952.970480] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.979439] env[63202]: DEBUG oslo_vmware.api [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385605, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.047175] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b00e946c-fc12-46a5-902a-7de0dddb8b64 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Lock "a961b2a0-39c1-4267-9229-068e2b6ecd67" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.249s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.227937] env[63202]: DEBUG oslo_concurrency.lockutils [req-c9e0b5a3-f18d-4aa8-b682-a9eda0f49f5a req-8896d125-0a60-441b-8c71-b928070cc396 service nova] Releasing lock "refresh_cache-2110dfa7-1795-4c8f-92bf-18921409c99f" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.256539] env[63202]: DEBUG nova.compute.manager [req-5f917de2-a2f8-4afa-a92b-999067db5500 req-b2f9a39c-dc00-44db-bae4-a0d029e1510d service nova] [instance: da285417-bd38-4387-8521-df326ca0e326] Received event network-vif-plugged-a0547ee8-b263-49d2-8152-cbac29b8c25b {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 953.256787] env[63202]: DEBUG oslo_concurrency.lockutils [req-5f917de2-a2f8-4afa-a92b-999067db5500 req-b2f9a39c-dc00-44db-bae4-a0d029e1510d service nova] Acquiring lock "da285417-bd38-4387-8521-df326ca0e326-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.257022] env[63202]: DEBUG oslo_concurrency.lockutils [req-5f917de2-a2f8-4afa-a92b-999067db5500 req-b2f9a39c-dc00-44db-bae4-a0d029e1510d service nova] Lock "da285417-bd38-4387-8521-df326ca0e326-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.257203] env[63202]: DEBUG oslo_concurrency.lockutils [req-5f917de2-a2f8-4afa-a92b-999067db5500 req-b2f9a39c-dc00-44db-bae4-a0d029e1510d service nova] Lock "da285417-bd38-4387-8521-df326ca0e326-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.257673] env[63202]: DEBUG nova.compute.manager [req-5f917de2-a2f8-4afa-a92b-999067db5500 req-b2f9a39c-dc00-44db-bae4-a0d029e1510d service nova] [instance: da285417-bd38-4387-8521-df326ca0e326] No waiting events found dispatching network-vif-plugged-a0547ee8-b263-49d2-8152-cbac29b8c25b {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 953.257673] env[63202]: WARNING nova.compute.manager [req-5f917de2-a2f8-4afa-a92b-999067db5500 req-b2f9a39c-dc00-44db-bae4-a0d029e1510d service nova] [instance: da285417-bd38-4387-8521-df326ca0e326] Received unexpected event network-vif-plugged-a0547ee8-b263-49d2-8152-cbac29b8c25b for instance with vm_state building and task_state spawning. [ 953.260440] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4eda49d7-5b02-4d51-99e9-f29b250eca15 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Lock "af2bedc2-28ee-4679-ae38-1cceb2af05d6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.279s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.431556] env[63202]: DEBUG nova.network.neutron [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: da285417-bd38-4387-8521-df326ca0e326] Successfully updated port: a0547ee8-b263-49d2-8152-cbac29b8c25b {{(pid=63202) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 953.447099] env[63202]: DEBUG oslo_vmware.api [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52775c09-20e1-e588-7c16-f6a321278364, 'name': SearchDatastore_Task, 'duration_secs': 0.03604} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.452711] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e075a2f3-e347-41dc-9acc-3dc1716f54fe {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.459935] env[63202]: DEBUG oslo_vmware.api [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 953.459935] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52d7e3a9-c5a4-f2f3-1676-05edd1945208" [ 953.459935] env[63202]: _type = "Task" [ 953.459935] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.472258] env[63202]: DEBUG oslo_vmware.api [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52d7e3a9-c5a4-f2f3-1676-05edd1945208, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.482267] env[63202]: DEBUG oslo_vmware.api [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385605, 'name': PowerOnVM_Task, 'duration_secs': 0.405318} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.485303] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 953.485596] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-ee8f35ee-7121-42a0-9ceb-0e15a8cff7e1 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Updating instance 'e775e5e1-521a-4fc7-80e6-bcb6a70516c5' progress to 100 {{(pid=63202) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 953.540633] env[63202]: DEBUG nova.objects.instance [None req-4e400d69-10f2-4327-9888-f1914bbefb0f tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lazy-loading 'flavor' on Instance uuid 8755bceb-d510-4429-bd98-d6a63faf739d {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 953.570480] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02cb7fcd-c4ea-4848-8612-d4ad2e52b799 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.580192] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48d63514-a1f2-4c1f-ab29-40921fd9a895 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.610860] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c2a9246-b6e6-446c-ab7e-fdb406d8f138 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.619498] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1c0955f-fc90-4f1e-b923-1428feb5cd88 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.633734] env[63202]: DEBUG nova.compute.provider_tree [None req-02dae3ed-6d57-4066-a061-72500d73ac65 tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 953.711938] env[63202]: DEBUG oslo_concurrency.lockutils [None req-854b8b38-28af-4f84-ac47-238da88ee0d4 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "8755bceb-d510-4429-bd98-d6a63faf739d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.941085] env[63202]: DEBUG oslo_concurrency.lockutils [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquiring lock "refresh_cache-da285417-bd38-4387-8521-df326ca0e326" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.941227] env[63202]: DEBUG oslo_concurrency.lockutils [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquired lock "refresh_cache-da285417-bd38-4387-8521-df326ca0e326" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.941380] env[63202]: DEBUG nova.network.neutron [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: da285417-bd38-4387-8521-df326ca0e326] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 953.974262] env[63202]: DEBUG oslo_vmware.api [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52d7e3a9-c5a4-f2f3-1676-05edd1945208, 'name': SearchDatastore_Task, 'duration_secs': 0.022744} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.974537] env[63202]: DEBUG oslo_concurrency.lockutils [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.974791] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] 2110dfa7-1795-4c8f-92bf-18921409c99f/2110dfa7-1795-4c8f-92bf-18921409c99f.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 953.975062] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6db17d25-7f74-4518-a62f-92b8dfe114b3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.983124] env[63202]: DEBUG oslo_vmware.api [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 953.983124] env[63202]: value = "task-1385606" [ 953.983124] env[63202]: _type = "Task" [ 953.983124] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.994818] env[63202]: DEBUG oslo_vmware.api [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385606, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.045743] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4e400d69-10f2-4327-9888-f1914bbefb0f tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "8755bceb-d510-4429-bd98-d6a63faf739d" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.782s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.047431] env[63202]: DEBUG oslo_concurrency.lockutils [None req-854b8b38-28af-4f84-ac47-238da88ee0d4 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "8755bceb-d510-4429-bd98-d6a63faf739d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.335s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.047701] env[63202]: DEBUG oslo_concurrency.lockutils [None req-854b8b38-28af-4f84-ac47-238da88ee0d4 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "8755bceb-d510-4429-bd98-d6a63faf739d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.050399] env[63202]: DEBUG oslo_concurrency.lockutils [None req-854b8b38-28af-4f84-ac47-238da88ee0d4 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "8755bceb-d510-4429-bd98-d6a63faf739d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.050399] env[63202]: DEBUG oslo_concurrency.lockutils [None req-854b8b38-28af-4f84-ac47-238da88ee0d4 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "8755bceb-d510-4429-bd98-d6a63faf739d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.050399] env[63202]: INFO nova.compute.manager [None req-854b8b38-28af-4f84-ac47-238da88ee0d4 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Terminating instance [ 954.055096] env[63202]: DEBUG nova.compute.manager [None req-854b8b38-28af-4f84-ac47-238da88ee0d4 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 954.055096] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-854b8b38-28af-4f84-ac47-238da88ee0d4 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 954.055096] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f1d5a1de-4319-4116-94af-0cd40c599e95 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.061673] env[63202]: DEBUG oslo_vmware.api [None req-854b8b38-28af-4f84-ac47-238da88ee0d4 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 954.061673] env[63202]: value = "task-1385607" [ 954.061673] env[63202]: _type = "Task" [ 954.061673] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.072674] env[63202]: DEBUG oslo_vmware.api [None req-854b8b38-28af-4f84-ac47-238da88ee0d4 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385607, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.136808] env[63202]: DEBUG nova.scheduler.client.report [None req-02dae3ed-6d57-4066-a061-72500d73ac65 tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 954.488050] env[63202]: DEBUG nova.network.neutron [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: da285417-bd38-4387-8521-df326ca0e326] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 954.496613] env[63202]: DEBUG oslo_vmware.api [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385606, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.576021] env[63202]: DEBUG oslo_vmware.api [None req-854b8b38-28af-4f84-ac47-238da88ee0d4 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385607, 'name': PowerOffVM_Task, 'duration_secs': 0.220673} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.576498] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-854b8b38-28af-4f84-ac47-238da88ee0d4 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 954.576866] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-854b8b38-28af-4f84-ac47-238da88ee0d4 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Volume detach. Driver type: vmdk {{(pid=63202) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 954.577172] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-854b8b38-28af-4f84-ac47-238da88ee0d4 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294173', 'volume_id': 'b7ce888f-0703-4bd3-89f9-198e09e4dc43', 'name': 'volume-b7ce888f-0703-4bd3-89f9-198e09e4dc43', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8755bceb-d510-4429-bd98-d6a63faf739d', 'attached_at': '', 'detached_at': '', 'volume_id': 'b7ce888f-0703-4bd3-89f9-198e09e4dc43', 'serial': 'b7ce888f-0703-4bd3-89f9-198e09e4dc43'} {{(pid=63202) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 954.581024] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8c74f04-d123-4faa-a656-4a7c964240ee {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.610243] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d60de347-8c54-47a4-8a95-b7fe2205dc28 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.623053] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fac97f62-caac-4d7e-8c52-bd14394d7037 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.653454] env[63202]: DEBUG oslo_concurrency.lockutils [None req-02dae3ed-6d57-4066-a061-72500d73ac65 tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.930s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.655949] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 10.251s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.656217] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.656422] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63202) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 954.656781] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d316d71b-590a-440a-b0fc-2e046c2315ac tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.271s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.658567] env[63202]: DEBUG nova.objects.instance [None req-d316d71b-590a-440a-b0fc-2e046c2315ac tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Lazy-loading 'resources' on Instance uuid d0d6e380-9337-4f69-8434-6a33ac8a33d6 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 954.661819] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9222a3dc-696c-4c89-80b8-6b0641ad0fd7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.666792] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-411f6d10-f005-4002-9f5e-de48e884aaea {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.687759] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e97a97b6-0750-428c-b60f-ca31c8587781 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.693141] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-854b8b38-28af-4f84-ac47-238da88ee0d4 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] The volume has not been displaced from its original location: [datastore1] volume-b7ce888f-0703-4bd3-89f9-198e09e4dc43/volume-b7ce888f-0703-4bd3-89f9-198e09e4dc43.vmdk. No consolidation needed. {{(pid=63202) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 954.699206] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-854b8b38-28af-4f84-ac47-238da88ee0d4 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Reconfiguring VM instance instance-0000004f to detach disk 2001 {{(pid=63202) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 954.704034] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8417fbd6-00c5-43b9-b26f-845293a436e1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.718945] env[63202]: INFO nova.scheduler.client.report [None req-02dae3ed-6d57-4066-a061-72500d73ac65 tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Deleted allocations for instance 2243bb69-0dc5-49cd-b94e-73e703cbadc0 [ 954.734460] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-559a4bdb-a1af-4336-81d5-9088fb64a8b5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.737352] env[63202]: DEBUG oslo_vmware.api [None req-854b8b38-28af-4f84-ac47-238da88ee0d4 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 954.737352] env[63202]: value = "task-1385608" [ 954.737352] env[63202]: _type = "Task" [ 954.737352] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.746134] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d17f40a6-f10a-42a3-9874-30f2017a514c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.754802] env[63202]: DEBUG oslo_vmware.api [None req-854b8b38-28af-4f84-ac47-238da88ee0d4 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385608, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.783621] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181089MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=63202) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 954.783849] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.785084] env[63202]: DEBUG nova.network.neutron [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: da285417-bd38-4387-8521-df326ca0e326] Updating instance_info_cache with network_info: [{"id": "a0547ee8-b263-49d2-8152-cbac29b8c25b", "address": "fa:16:3e:90:ea:09", "network": {"id": "e6c760f6-ded9-4905-b303-dbcde0d037b9", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1433567753-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e914f87450949b1a39866e8cfa3a0eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "26472e27-9835-4f87-ab7f-ca24dfee4e83", "external-id": "nsx-vlan-transportzone-335", "segmentation_id": 335, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0547ee8-b2", "ovs_interfaceid": "a0547ee8-b263-49d2-8152-cbac29b8c25b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.996545] env[63202]: DEBUG oslo_vmware.api [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385606, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.522337} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.996811] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] 2110dfa7-1795-4c8f-92bf-18921409c99f/2110dfa7-1795-4c8f-92bf-18921409c99f.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 954.997042] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 954.997317] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1c3305c9-fee8-4ed8-8351-31d18dc9589b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.015467] env[63202]: DEBUG oslo_vmware.api [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 955.015467] env[63202]: value = "task-1385609" [ 955.015467] env[63202]: _type = "Task" [ 955.015467] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.029230] env[63202]: DEBUG oslo_vmware.api [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385609, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.244309] env[63202]: DEBUG oslo_concurrency.lockutils [None req-02dae3ed-6d57-4066-a061-72500d73ac65 tempest-ServersTestJSON-483277806 tempest-ServersTestJSON-483277806-project-member] Lock "2243bb69-0dc5-49cd-b94e-73e703cbadc0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.483s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.266635] env[63202]: DEBUG oslo_vmware.api [None req-854b8b38-28af-4f84-ac47-238da88ee0d4 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385608, 'name': ReconfigVM_Task, 'duration_secs': 0.4} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.266635] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-854b8b38-28af-4f84-ac47-238da88ee0d4 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Reconfigured VM instance instance-0000004f to detach disk 2001 {{(pid=63202) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 955.269677] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-58941eb3-168d-419d-82c0-8f642dbfa5f6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.288947] env[63202]: DEBUG oslo_concurrency.lockutils [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Releasing lock "refresh_cache-da285417-bd38-4387-8521-df326ca0e326" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.290335] env[63202]: DEBUG nova.compute.manager [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: da285417-bd38-4387-8521-df326ca0e326] Instance network_info: |[{"id": "a0547ee8-b263-49d2-8152-cbac29b8c25b", "address": "fa:16:3e:90:ea:09", "network": {"id": "e6c760f6-ded9-4905-b303-dbcde0d037b9", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1433567753-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e914f87450949b1a39866e8cfa3a0eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "26472e27-9835-4f87-ab7f-ca24dfee4e83", "external-id": "nsx-vlan-transportzone-335", "segmentation_id": 335, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0547ee8-b2", "ovs_interfaceid": "a0547ee8-b263-49d2-8152-cbac29b8c25b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 955.290567] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: da285417-bd38-4387-8521-df326ca0e326] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:90:ea:09', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '26472e27-9835-4f87-ab7f-ca24dfee4e83', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a0547ee8-b263-49d2-8152-cbac29b8c25b', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 955.302839] env[63202]: DEBUG oslo.service.loopingcall [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 955.308491] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: da285417-bd38-4387-8521-df326ca0e326] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 955.308491] env[63202]: DEBUG oslo_vmware.api [None req-854b8b38-28af-4f84-ac47-238da88ee0d4 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 955.308491] env[63202]: value = "task-1385610" [ 955.308491] env[63202]: _type = "Task" [ 955.308491] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.308491] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9782a663-8d5c-4e02-aabc-83a520741f02 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.342553] env[63202]: DEBUG oslo_vmware.api [None req-854b8b38-28af-4f84-ac47-238da88ee0d4 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385610, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.344108] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 955.344108] env[63202]: value = "task-1385611" [ 955.344108] env[63202]: _type = "Task" [ 955.344108] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.363039] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385611, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.480629] env[63202]: DEBUG nova.compute.manager [req-1ff9f17a-0d6b-4353-9cb5-3607f1b4e4cd req-97cd1bbf-5201-41b7-a61a-b0da3184706c service nova] [instance: da285417-bd38-4387-8521-df326ca0e326] Received event network-changed-a0547ee8-b263-49d2-8152-cbac29b8c25b {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 955.480829] env[63202]: DEBUG nova.compute.manager [req-1ff9f17a-0d6b-4353-9cb5-3607f1b4e4cd req-97cd1bbf-5201-41b7-a61a-b0da3184706c service nova] [instance: da285417-bd38-4387-8521-df326ca0e326] Refreshing instance network info cache due to event network-changed-a0547ee8-b263-49d2-8152-cbac29b8c25b. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 955.481178] env[63202]: DEBUG oslo_concurrency.lockutils [req-1ff9f17a-0d6b-4353-9cb5-3607f1b4e4cd req-97cd1bbf-5201-41b7-a61a-b0da3184706c service nova] Acquiring lock "refresh_cache-da285417-bd38-4387-8521-df326ca0e326" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.481376] env[63202]: DEBUG oslo_concurrency.lockutils [req-1ff9f17a-0d6b-4353-9cb5-3607f1b4e4cd req-97cd1bbf-5201-41b7-a61a-b0da3184706c service nova] Acquired lock "refresh_cache-da285417-bd38-4387-8521-df326ca0e326" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.481807] env[63202]: DEBUG nova.network.neutron [req-1ff9f17a-0d6b-4353-9cb5-3607f1b4e4cd req-97cd1bbf-5201-41b7-a61a-b0da3184706c service nova] [instance: da285417-bd38-4387-8521-df326ca0e326] Refreshing network info cache for port a0547ee8-b263-49d2-8152-cbac29b8c25b {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 955.529524] env[63202]: DEBUG oslo_vmware.api [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385609, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.2482} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.529524] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 955.530231] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-475b02f3-bd99-4f70-96c4-b3706b8b7a72 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.534995] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b1af270-6d7d-4201-ad29-26b445926eb3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.553123] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d48c105-85d4-4cb7-ad69-5f847d9c4907 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.569668] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Reconfiguring VM instance instance-00000053 to attach disk [datastore1] 2110dfa7-1795-4c8f-92bf-18921409c99f/2110dfa7-1795-4c8f-92bf-18921409c99f.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 955.574027] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-570f59cc-b844-4f7b-8e0b-d498dbf68dde {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.617875] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-725d897b-e5a9-4c8a-9263-88368e15e56e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.620918] env[63202]: DEBUG oslo_vmware.api [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 955.620918] env[63202]: value = "task-1385612" [ 955.620918] env[63202]: _type = "Task" [ 955.620918] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.628676] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f4f489f-0b57-4990-874b-aa650607a808 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.636557] env[63202]: DEBUG oslo_vmware.api [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385612, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.652026] env[63202]: DEBUG nova.compute.provider_tree [None req-d316d71b-590a-440a-b0fc-2e046c2315ac tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 955.834904] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d857d01b-4657-47db-afda-873bd808eb35 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "e775e5e1-521a-4fc7-80e6-bcb6a70516c5" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.835603] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d857d01b-4657-47db-afda-873bd808eb35 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "e775e5e1-521a-4fc7-80e6-bcb6a70516c5" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.838018] env[63202]: DEBUG nova.compute.manager [None req-d857d01b-4657-47db-afda-873bd808eb35 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Going to confirm migration 1 {{(pid=63202) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 955.841251] env[63202]: DEBUG oslo_vmware.api [None req-854b8b38-28af-4f84-ac47-238da88ee0d4 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385610, 'name': ReconfigVM_Task, 'duration_secs': 0.263812} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.842120] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-854b8b38-28af-4f84-ac47-238da88ee0d4 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294173', 'volume_id': 'b7ce888f-0703-4bd3-89f9-198e09e4dc43', 'name': 'volume-b7ce888f-0703-4bd3-89f9-198e09e4dc43', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8755bceb-d510-4429-bd98-d6a63faf739d', 'attached_at': '', 'detached_at': '', 'volume_id': 'b7ce888f-0703-4bd3-89f9-198e09e4dc43', 'serial': 'b7ce888f-0703-4bd3-89f9-198e09e4dc43'} {{(pid=63202) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 955.845020] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-854b8b38-28af-4f84-ac47-238da88ee0d4 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 955.845020] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37f7a84a-03fd-4138-9ab2-18a0841b0eea {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.859894] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-854b8b38-28af-4f84-ac47-238da88ee0d4 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 955.863403] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f5574258-9715-4377-ac55-3293a1036319 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.865628] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385611, 'name': CreateVM_Task} progress is 99%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.867328] env[63202]: DEBUG nova.compute.manager [None req-889ec7f6-4453-4d63-bd3b-38aa55b11694 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 955.871047] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f363a8b7-eeb6-4084-8338-d152b9c90472 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.979904] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-854b8b38-28af-4f84-ac47-238da88ee0d4 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 955.980143] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-854b8b38-28af-4f84-ac47-238da88ee0d4 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Deleting contents of the VM from datastore datastore2 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 955.980379] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-854b8b38-28af-4f84-ac47-238da88ee0d4 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Deleting the datastore file [datastore2] 8755bceb-d510-4429-bd98-d6a63faf739d {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 955.980675] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-555c6a76-ee03-4d6e-8aa9-51aca7e6b2a2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.989224] env[63202]: DEBUG oslo_vmware.api [None req-854b8b38-28af-4f84-ac47-238da88ee0d4 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 955.989224] env[63202]: value = "task-1385614" [ 955.989224] env[63202]: _type = "Task" [ 955.989224] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.004392] env[63202]: DEBUG oslo_vmware.api [None req-854b8b38-28af-4f84-ac47-238da88ee0d4 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385614, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.134124] env[63202]: DEBUG oslo_vmware.api [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385612, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.151855] env[63202]: DEBUG nova.scheduler.client.report [None req-d316d71b-590a-440a-b0fc-2e046c2315ac tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 956.328229] env[63202]: DEBUG nova.network.neutron [req-1ff9f17a-0d6b-4353-9cb5-3607f1b4e4cd req-97cd1bbf-5201-41b7-a61a-b0da3184706c service nova] [instance: da285417-bd38-4387-8521-df326ca0e326] Updated VIF entry in instance network info cache for port a0547ee8-b263-49d2-8152-cbac29b8c25b. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 956.328475] env[63202]: DEBUG nova.network.neutron [req-1ff9f17a-0d6b-4353-9cb5-3607f1b4e4cd req-97cd1bbf-5201-41b7-a61a-b0da3184706c service nova] [instance: da285417-bd38-4387-8521-df326ca0e326] Updating instance_info_cache with network_info: [{"id": "a0547ee8-b263-49d2-8152-cbac29b8c25b", "address": "fa:16:3e:90:ea:09", "network": {"id": "e6c760f6-ded9-4905-b303-dbcde0d037b9", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1433567753-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e914f87450949b1a39866e8cfa3a0eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "26472e27-9835-4f87-ab7f-ca24dfee4e83", "external-id": "nsx-vlan-transportzone-335", "segmentation_id": 335, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0547ee8-b2", "ovs_interfaceid": "a0547ee8-b263-49d2-8152-cbac29b8c25b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.363801] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385611, 'name': CreateVM_Task} progress is 99%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.385210] env[63202]: INFO nova.compute.manager [None req-889ec7f6-4453-4d63-bd3b-38aa55b11694 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] instance snapshotting [ 956.389629] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c6ea57e-ccb3-431d-9fb0-8150f0e892c6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.420437] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a336f51a-284b-4478-93c0-cf064f58d697 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.501287] env[63202]: DEBUG oslo_vmware.api [None req-854b8b38-28af-4f84-ac47-238da88ee0d4 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385614, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.44609} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.501661] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-854b8b38-28af-4f84-ac47-238da88ee0d4 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 956.501856] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-854b8b38-28af-4f84-ac47-238da88ee0d4 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Deleted contents of the VM from datastore datastore2 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 956.502172] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-854b8b38-28af-4f84-ac47-238da88ee0d4 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 956.502765] env[63202]: INFO nova.compute.manager [None req-854b8b38-28af-4f84-ac47-238da88ee0d4 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Took 2.45 seconds to destroy the instance on the hypervisor. [ 956.502843] env[63202]: DEBUG oslo.service.loopingcall [None req-854b8b38-28af-4f84-ac47-238da88ee0d4 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 956.503256] env[63202]: DEBUG nova.compute.manager [-] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 956.503763] env[63202]: DEBUG nova.network.neutron [-] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 956.506646] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d857d01b-4657-47db-afda-873bd808eb35 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "refresh_cache-e775e5e1-521a-4fc7-80e6-bcb6a70516c5" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.507129] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d857d01b-4657-47db-afda-873bd808eb35 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquired lock "refresh_cache-e775e5e1-521a-4fc7-80e6-bcb6a70516c5" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.507129] env[63202]: DEBUG nova.network.neutron [None req-d857d01b-4657-47db-afda-873bd808eb35 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 956.507260] env[63202]: DEBUG nova.objects.instance [None req-d857d01b-4657-47db-afda-873bd808eb35 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lazy-loading 'info_cache' on Instance uuid e775e5e1-521a-4fc7-80e6-bcb6a70516c5 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 956.586648] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dcb77ba6-82bd-49fa-bf90-c76e177d7388 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Acquiring lock "b1bccea9-2d79-431a-8be0-0a5ab293542a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.586648] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dcb77ba6-82bd-49fa-bf90-c76e177d7388 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Lock "b1bccea9-2d79-431a-8be0-0a5ab293542a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.586648] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dcb77ba6-82bd-49fa-bf90-c76e177d7388 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Acquiring lock "b1bccea9-2d79-431a-8be0-0a5ab293542a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.586648] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dcb77ba6-82bd-49fa-bf90-c76e177d7388 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Lock "b1bccea9-2d79-431a-8be0-0a5ab293542a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.587194] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dcb77ba6-82bd-49fa-bf90-c76e177d7388 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Lock "b1bccea9-2d79-431a-8be0-0a5ab293542a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.589778] env[63202]: INFO nova.compute.manager [None req-dcb77ba6-82bd-49fa-bf90-c76e177d7388 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Terminating instance [ 956.592037] env[63202]: DEBUG nova.compute.manager [None req-dcb77ba6-82bd-49fa-bf90-c76e177d7388 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 956.592765] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-dcb77ba6-82bd-49fa-bf90-c76e177d7388 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 956.593223] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5847f961-c77b-4741-b630-7276fdb64846 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.604399] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcb77ba6-82bd-49fa-bf90-c76e177d7388 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 956.604707] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-29e4bc31-b3b3-47ff-9c6f-dbb956b56bb2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.613212] env[63202]: DEBUG oslo_vmware.api [None req-dcb77ba6-82bd-49fa-bf90-c76e177d7388 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Waiting for the task: (returnval){ [ 956.613212] env[63202]: value = "task-1385615" [ 956.613212] env[63202]: _type = "Task" [ 956.613212] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.622837] env[63202]: DEBUG oslo_vmware.api [None req-dcb77ba6-82bd-49fa-bf90-c76e177d7388 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385615, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.633139] env[63202]: DEBUG oslo_vmware.api [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385612, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.665562] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d316d71b-590a-440a-b0fc-2e046c2315ac tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.009s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.668634] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 9.082s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.668748] env[63202]: DEBUG nova.objects.instance [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63202) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 956.671713] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7e26c414-ad5f-4538-9c98-9fcc1de5bbcb tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Acquiring lock "f3652744-e072-4700-80d4-b9eca414c5cb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.675404] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7e26c414-ad5f-4538-9c98-9fcc1de5bbcb tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Lock "f3652744-e072-4700-80d4-b9eca414c5cb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.004s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.675819] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7e26c414-ad5f-4538-9c98-9fcc1de5bbcb tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Acquiring lock "f3652744-e072-4700-80d4-b9eca414c5cb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.675910] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7e26c414-ad5f-4538-9c98-9fcc1de5bbcb tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Lock "f3652744-e072-4700-80d4-b9eca414c5cb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.676091] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7e26c414-ad5f-4538-9c98-9fcc1de5bbcb tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Lock "f3652744-e072-4700-80d4-b9eca414c5cb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.681686] env[63202]: INFO nova.compute.manager [None req-7e26c414-ad5f-4538-9c98-9fcc1de5bbcb tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Terminating instance [ 956.687057] env[63202]: DEBUG nova.compute.manager [None req-7e26c414-ad5f-4538-9c98-9fcc1de5bbcb tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 956.687286] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-7e26c414-ad5f-4538-9c98-9fcc1de5bbcb tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 956.688178] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b481605-f85a-438c-9bd1-86c7aa634ed8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.692235] env[63202]: INFO nova.scheduler.client.report [None req-d316d71b-590a-440a-b0fc-2e046c2315ac tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Deleted allocations for instance d0d6e380-9337-4f69-8434-6a33ac8a33d6 [ 956.702102] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e26c414-ad5f-4538-9c98-9fcc1de5bbcb tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 956.702372] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-778c4b86-a5e2-4925-9ca5-49b734fc5743 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.710901] env[63202]: DEBUG oslo_vmware.api [None req-7e26c414-ad5f-4538-9c98-9fcc1de5bbcb tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Waiting for the task: (returnval){ [ 956.710901] env[63202]: value = "task-1385616" [ 956.710901] env[63202]: _type = "Task" [ 956.710901] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.725267] env[63202]: DEBUG oslo_vmware.api [None req-7e26c414-ad5f-4538-9c98-9fcc1de5bbcb tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385616, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.832946] env[63202]: DEBUG oslo_concurrency.lockutils [req-1ff9f17a-0d6b-4353-9cb5-3607f1b4e4cd req-97cd1bbf-5201-41b7-a61a-b0da3184706c service nova] Releasing lock "refresh_cache-da285417-bd38-4387-8521-df326ca0e326" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.864972] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385611, 'name': CreateVM_Task, 'duration_secs': 1.514908} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.865283] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: da285417-bd38-4387-8521-df326ca0e326] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 956.866505] env[63202]: DEBUG oslo_concurrency.lockutils [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/376aee6c-1f98-4a60-858b-40c55f65b90e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.866768] env[63202]: DEBUG oslo_concurrency.lockutils [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquired lock "[datastore1] devstack-image-cache_base/376aee6c-1f98-4a60-858b-40c55f65b90e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.867211] env[63202]: DEBUG oslo_concurrency.lockutils [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/376aee6c-1f98-4a60-858b-40c55f65b90e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 956.867545] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c25d1ec7-168b-4a11-843c-dfea82e927f2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.875509] env[63202]: DEBUG oslo_vmware.api [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for the task: (returnval){ [ 956.875509] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52499ee5-9619-2b0c-2e90-ffdcc0811dec" [ 956.875509] env[63202]: _type = "Task" [ 956.875509] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.889622] env[63202]: DEBUG oslo_vmware.api [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52499ee5-9619-2b0c-2e90-ffdcc0811dec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.939129] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-889ec7f6-4453-4d63-bd3b-38aa55b11694 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Creating Snapshot of the VM instance {{(pid=63202) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 956.939659] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-013672e4-f4cf-407c-8166-50e662ff86b5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.950533] env[63202]: DEBUG oslo_vmware.api [None req-889ec7f6-4453-4d63-bd3b-38aa55b11694 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Waiting for the task: (returnval){ [ 956.950533] env[63202]: value = "task-1385617" [ 956.950533] env[63202]: _type = "Task" [ 956.950533] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.966343] env[63202]: DEBUG oslo_vmware.api [None req-889ec7f6-4453-4d63-bd3b-38aa55b11694 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385617, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.127163] env[63202]: DEBUG oslo_vmware.api [None req-dcb77ba6-82bd-49fa-bf90-c76e177d7388 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385615, 'name': PowerOffVM_Task, 'duration_secs': 0.353266} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.129031] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcb77ba6-82bd-49fa-bf90-c76e177d7388 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 957.131900] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-dcb77ba6-82bd-49fa-bf90-c76e177d7388 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 957.131900] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f5528ac3-35b0-4a14-b0de-23429727b3ff {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.141022] env[63202]: DEBUG oslo_vmware.api [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385612, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.201642] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d316d71b-590a-440a-b0fc-2e046c2315ac tempest-ServerPasswordTestJSON-989293939 tempest-ServerPasswordTestJSON-989293939-project-member] Lock "d0d6e380-9337-4f69-8434-6a33ac8a33d6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.319s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.211752] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-dcb77ba6-82bd-49fa-bf90-c76e177d7388 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 957.212039] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-dcb77ba6-82bd-49fa-bf90-c76e177d7388 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Deleting contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 957.212266] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-dcb77ba6-82bd-49fa-bf90-c76e177d7388 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Deleting the datastore file [datastore1] b1bccea9-2d79-431a-8be0-0a5ab293542a {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 957.212675] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dc2630bf-7487-4f3b-95bf-4013431640b7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.229855] env[63202]: DEBUG oslo_vmware.api [None req-7e26c414-ad5f-4538-9c98-9fcc1de5bbcb tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385616, 'name': PowerOffVM_Task, 'duration_secs': 0.249872} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.231585] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e26c414-ad5f-4538-9c98-9fcc1de5bbcb tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 957.231772] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-7e26c414-ad5f-4538-9c98-9fcc1de5bbcb tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 957.233963] env[63202]: DEBUG oslo_vmware.api [None req-dcb77ba6-82bd-49fa-bf90-c76e177d7388 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Waiting for the task: (returnval){ [ 957.233963] env[63202]: value = "task-1385619" [ 957.233963] env[63202]: _type = "Task" [ 957.233963] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.233963] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4462490e-b651-4a3d-a432-582c2f0dcf81 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.244873] env[63202]: DEBUG oslo_vmware.api [None req-dcb77ba6-82bd-49fa-bf90-c76e177d7388 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385619, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.309347] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-7e26c414-ad5f-4538-9c98-9fcc1de5bbcb tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 957.309625] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-7e26c414-ad5f-4538-9c98-9fcc1de5bbcb tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Deleting contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 957.309846] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e26c414-ad5f-4538-9c98-9fcc1de5bbcb tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Deleting the datastore file [datastore1] f3652744-e072-4700-80d4-b9eca414c5cb {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 957.310162] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0f6067a0-4b78-483b-aa57-07b858506b64 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.315896] env[63202]: DEBUG nova.network.neutron [-] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.318693] env[63202]: DEBUG oslo_vmware.api [None req-7e26c414-ad5f-4538-9c98-9fcc1de5bbcb tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Waiting for the task: (returnval){ [ 957.318693] env[63202]: value = "task-1385621" [ 957.318693] env[63202]: _type = "Task" [ 957.318693] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.327974] env[63202]: DEBUG oslo_vmware.api [None req-7e26c414-ad5f-4538-9c98-9fcc1de5bbcb tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385621, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.386667] env[63202]: DEBUG oslo_concurrency.lockutils [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Releasing lock "[datastore1] devstack-image-cache_base/376aee6c-1f98-4a60-858b-40c55f65b90e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.387077] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: da285417-bd38-4387-8521-df326ca0e326] Processing image 376aee6c-1f98-4a60-858b-40c55f65b90e {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 957.387245] env[63202]: DEBUG oslo_concurrency.lockutils [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/376aee6c-1f98-4a60-858b-40c55f65b90e/376aee6c-1f98-4a60-858b-40c55f65b90e.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.387365] env[63202]: DEBUG oslo_concurrency.lockutils [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquired lock "[datastore1] devstack-image-cache_base/376aee6c-1f98-4a60-858b-40c55f65b90e/376aee6c-1f98-4a60-858b-40c55f65b90e.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.387760] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 957.387968] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bee02cd3-edad-4292-9804-7f2b004635f4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.398593] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 957.398884] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 957.399959] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ef287e9-226b-4182-a927-e97d6c3f82fe {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.406815] env[63202]: DEBUG oslo_vmware.api [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for the task: (returnval){ [ 957.406815] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52d5b0d5-07a8-8357-4a2a-968769bfd76d" [ 957.406815] env[63202]: _type = "Task" [ 957.406815] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.416134] env[63202]: DEBUG oslo_vmware.api [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52d5b0d5-07a8-8357-4a2a-968769bfd76d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.462126] env[63202]: DEBUG oslo_vmware.api [None req-889ec7f6-4453-4d63-bd3b-38aa55b11694 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385617, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.567937] env[63202]: DEBUG nova.compute.manager [req-0436676c-77a9-43db-a9dc-2b0f6c06ce30 req-f3f73826-a898-4a9f-b059-22b2e23bf738 service nova] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Received event network-vif-deleted-faa30f43-2e38-4dbb-97ca-1b4c17565310 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 957.638697] env[63202]: DEBUG oslo_vmware.api [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385612, 'name': ReconfigVM_Task, 'duration_secs': 1.581392} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.639334] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Reconfigured VM instance instance-00000053 to attach disk [datastore1] 2110dfa7-1795-4c8f-92bf-18921409c99f/2110dfa7-1795-4c8f-92bf-18921409c99f.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 957.640906] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9bce39de-cbb7-4df6-b4f7-621400170997 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.652143] env[63202]: DEBUG oslo_vmware.api [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 957.652143] env[63202]: value = "task-1385622" [ 957.652143] env[63202]: _type = "Task" [ 957.652143] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.663987] env[63202]: DEBUG oslo_concurrency.lockutils [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquiring lock "8437f856-a707-49c5-b8eb-5a22cdb990f8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.664454] env[63202]: DEBUG oslo_concurrency.lockutils [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lock "8437f856-a707-49c5-b8eb-5a22cdb990f8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.671296] env[63202]: DEBUG oslo_vmware.api [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385622, 'name': Rename_Task} progress is 6%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.685485] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8643caef-ac42-4a24-a3c7-e53c7f973961 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.017s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.687805] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7de759cc-c39d-4058-8625-09664c11b490 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.021s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.688188] env[63202]: DEBUG nova.objects.instance [None req-7de759cc-c39d-4058-8625-09664c11b490 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Lazy-loading 'resources' on Instance uuid 3d1518d3-bed2-4b7b-af1d-d7da49e92874 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 957.745393] env[63202]: DEBUG oslo_vmware.api [None req-dcb77ba6-82bd-49fa-bf90-c76e177d7388 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385619, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.171452} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.745871] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-dcb77ba6-82bd-49fa-bf90-c76e177d7388 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 957.749232] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-dcb77ba6-82bd-49fa-bf90-c76e177d7388 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Deleted contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 957.749232] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-dcb77ba6-82bd-49fa-bf90-c76e177d7388 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 957.749232] env[63202]: INFO nova.compute.manager [None req-dcb77ba6-82bd-49fa-bf90-c76e177d7388 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Took 1.15 seconds to destroy the instance on the hypervisor. [ 957.749232] env[63202]: DEBUG oslo.service.loopingcall [None req-dcb77ba6-82bd-49fa-bf90-c76e177d7388 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 957.749232] env[63202]: DEBUG nova.compute.manager [-] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 957.749232] env[63202]: DEBUG nova.network.neutron [-] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 957.823023] env[63202]: INFO nova.compute.manager [-] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Took 1.32 seconds to deallocate network for instance. [ 957.836734] env[63202]: DEBUG oslo_vmware.api [None req-7e26c414-ad5f-4538-9c98-9fcc1de5bbcb tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Task: {'id': task-1385621, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.166029} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.836734] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e26c414-ad5f-4538-9c98-9fcc1de5bbcb tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 957.836734] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-7e26c414-ad5f-4538-9c98-9fcc1de5bbcb tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Deleted contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 957.836734] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-7e26c414-ad5f-4538-9c98-9fcc1de5bbcb tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 957.836734] env[63202]: INFO nova.compute.manager [None req-7e26c414-ad5f-4538-9c98-9fcc1de5bbcb tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Took 1.15 seconds to destroy the instance on the hypervisor. [ 957.837072] env[63202]: DEBUG oslo.service.loopingcall [None req-7e26c414-ad5f-4538-9c98-9fcc1de5bbcb tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 957.837072] env[63202]: DEBUG nova.compute.manager [-] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 957.837072] env[63202]: DEBUG nova.network.neutron [-] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 957.917927] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: da285417-bd38-4387-8521-df326ca0e326] Preparing fetch location {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 957.918227] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: da285417-bd38-4387-8521-df326ca0e326] Fetch image to [datastore1] OSTACK_IMG_8b975109-76f2-4f78-a4b5-029f151227fc/OSTACK_IMG_8b975109-76f2-4f78-a4b5-029f151227fc.vmdk {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 957.918439] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: da285417-bd38-4387-8521-df326ca0e326] Downloading stream optimized image 376aee6c-1f98-4a60-858b-40c55f65b90e to [datastore1] OSTACK_IMG_8b975109-76f2-4f78-a4b5-029f151227fc/OSTACK_IMG_8b975109-76f2-4f78-a4b5-029f151227fc.vmdk on the data store datastore1 as vApp {{(pid=63202) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 957.918629] env[63202]: DEBUG nova.virt.vmwareapi.images [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: da285417-bd38-4387-8521-df326ca0e326] Downloading image file data 376aee6c-1f98-4a60-858b-40c55f65b90e to the ESX as VM named 'OSTACK_IMG_8b975109-76f2-4f78-a4b5-029f151227fc' {{(pid=63202) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 957.977152] env[63202]: DEBUG nova.network.neutron [None req-d857d01b-4657-47db-afda-873bd808eb35 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Updating instance_info_cache with network_info: [{"id": "c7da6929-1bd1-4e2b-bcb8-609863cf6861", "address": "fa:16:3e:cf:6c:83", "network": {"id": "18390479-16ce-4012-9ba6-abf19b5616d1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1376026898-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22352c25bca8416a948014391a5389ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "085fb0ff-9285-4f1d-a008-a14da4844357", "external-id": "nsx-vlan-transportzone-729", "segmentation_id": 729, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7da6929-1b", "ovs_interfaceid": "c7da6929-1bd1-4e2b-bcb8-609863cf6861", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.981598] env[63202]: DEBUG oslo_vmware.api [None req-889ec7f6-4453-4d63-bd3b-38aa55b11694 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385617, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.020605] env[63202]: DEBUG oslo_vmware.rw_handles [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 958.020605] env[63202]: value = "resgroup-9" [ 958.020605] env[63202]: _type = "ResourcePool" [ 958.020605] env[63202]: }. {{(pid=63202) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 958.021047] env[63202]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-b4a3c758-5ddb-44a3-8f8b-58f35e39e310 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.046902] env[63202]: DEBUG oslo_vmware.rw_handles [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lease: (returnval){ [ 958.046902] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52899f9a-b78c-208d-e9eb-7c975e3bc8d4" [ 958.046902] env[63202]: _type = "HttpNfcLease" [ 958.046902] env[63202]: } obtained for vApp import into resource pool (val){ [ 958.046902] env[63202]: value = "resgroup-9" [ 958.046902] env[63202]: _type = "ResourcePool" [ 958.046902] env[63202]: }. {{(pid=63202) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 958.047312] env[63202]: DEBUG oslo_vmware.api [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for the lease: (returnval){ [ 958.047312] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52899f9a-b78c-208d-e9eb-7c975e3bc8d4" [ 958.047312] env[63202]: _type = "HttpNfcLease" [ 958.047312] env[63202]: } to be ready. {{(pid=63202) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 958.056172] env[63202]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 958.056172] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52899f9a-b78c-208d-e9eb-7c975e3bc8d4" [ 958.056172] env[63202]: _type = "HttpNfcLease" [ 958.056172] env[63202]: } is initializing. {{(pid=63202) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 958.165035] env[63202]: DEBUG oslo_vmware.api [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385622, 'name': Rename_Task, 'duration_secs': 0.234203} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.165551] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 958.165660] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bc15f539-1728-455c-bf7c-849ac6e707f1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.171018] env[63202]: DEBUG nova.compute.manager [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 958.180759] env[63202]: DEBUG oslo_vmware.api [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 958.180759] env[63202]: value = "task-1385624" [ 958.180759] env[63202]: _type = "Task" [ 958.180759] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.204182] env[63202]: DEBUG oslo_vmware.api [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385624, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.400650] env[63202]: INFO nova.compute.manager [None req-854b8b38-28af-4f84-ac47-238da88ee0d4 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Took 0.58 seconds to detach 1 volumes for instance. [ 958.446758] env[63202]: DEBUG nova.compute.manager [req-6004caee-e999-4b9c-8e7c-82c03fe0fdfb req-8f838d3d-a993-4671-a343-611e96fb41ac service nova] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Received event network-vif-deleted-79b59533-e84a-4ace-a615-596c8734838c {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 958.447049] env[63202]: INFO nova.compute.manager [req-6004caee-e999-4b9c-8e7c-82c03fe0fdfb req-8f838d3d-a993-4671-a343-611e96fb41ac service nova] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Neutron deleted interface 79b59533-e84a-4ace-a615-596c8734838c; detaching it from the instance and deleting it from the info cache [ 958.447237] env[63202]: DEBUG nova.network.neutron [req-6004caee-e999-4b9c-8e7c-82c03fe0fdfb req-8f838d3d-a993-4671-a343-611e96fb41ac service nova] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.484547] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d857d01b-4657-47db-afda-873bd808eb35 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Releasing lock "refresh_cache-e775e5e1-521a-4fc7-80e6-bcb6a70516c5" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.487019] env[63202]: DEBUG nova.objects.instance [None req-d857d01b-4657-47db-afda-873bd808eb35 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lazy-loading 'migration_context' on Instance uuid e775e5e1-521a-4fc7-80e6-bcb6a70516c5 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 958.495648] env[63202]: DEBUG oslo_vmware.api [None req-889ec7f6-4453-4d63-bd3b-38aa55b11694 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385617, 'name': CreateSnapshot_Task, 'duration_secs': 1.130103} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.496896] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-889ec7f6-4453-4d63-bd3b-38aa55b11694 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Created Snapshot of the VM instance {{(pid=63202) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 958.497654] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ce6c3cf-fd0b-4da4-9d89-cf86e43ce94b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.501444] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad19074c-1ac3-4959-8a3b-5ec1ba41f30f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.516090] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35d83934-9a82-4879-91c1-385994c4a9e9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.570294] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aa48952-c600-41b6-9d31-1abd093a4b9a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.582679] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba0b83d4-e518-4696-ba54-320621d9c6b6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.587273] env[63202]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 958.587273] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52899f9a-b78c-208d-e9eb-7c975e3bc8d4" [ 958.587273] env[63202]: _type = "HttpNfcLease" [ 958.587273] env[63202]: } is initializing. {{(pid=63202) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 958.601596] env[63202]: DEBUG nova.compute.provider_tree [None req-7de759cc-c39d-4058-8625-09664c11b490 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Updating inventory in ProviderTree for provider 79b33d17-3e75-494c-a550-67b275de2079 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 958.692219] env[63202]: DEBUG oslo_vmware.api [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385624, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.693419] env[63202]: DEBUG oslo_concurrency.lockutils [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.739747] env[63202]: DEBUG nova.network.neutron [-] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.793456] env[63202]: DEBUG nova.network.neutron [-] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.909973] env[63202]: DEBUG oslo_concurrency.lockutils [None req-854b8b38-28af-4f84-ac47-238da88ee0d4 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.952723] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-57b1af91-9ccb-4bce-8ce4-47ec2576ceba {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.960871] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb90b88f-1edf-4a12-8482-86919ffb32fd {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.994697] env[63202]: DEBUG nova.objects.base [None req-d857d01b-4657-47db-afda-873bd808eb35 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=63202) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 958.995072] env[63202]: DEBUG nova.compute.manager [req-6004caee-e999-4b9c-8e7c-82c03fe0fdfb req-8f838d3d-a993-4671-a343-611e96fb41ac service nova] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Detach interface failed, port_id=79b59533-e84a-4ace-a615-596c8734838c, reason: Instance f3652744-e072-4700-80d4-b9eca414c5cb could not be found. {{(pid=63202) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 958.996081] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfd9e9d1-cc5f-41bb-b72c-eaa142eb6fed {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.018606] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-353ff32c-4689-45ef-9d74-3df5c04ba2db {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.030548] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-889ec7f6-4453-4d63-bd3b-38aa55b11694 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Creating linked-clone VM from snapshot {{(pid=63202) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 959.032321] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-ad86ac71-c5de-4617-9b5e-f8096cf2c32f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.035342] env[63202]: DEBUG oslo_vmware.api [None req-d857d01b-4657-47db-afda-873bd808eb35 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for the task: (returnval){ [ 959.035342] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52f3cf4a-ab37-e791-e9aa-67bc02a3c5d2" [ 959.035342] env[63202]: _type = "Task" [ 959.035342] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.042890] env[63202]: DEBUG oslo_vmware.api [None req-889ec7f6-4453-4d63-bd3b-38aa55b11694 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Waiting for the task: (returnval){ [ 959.042890] env[63202]: value = "task-1385625" [ 959.042890] env[63202]: _type = "Task" [ 959.042890] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.047530] env[63202]: DEBUG oslo_vmware.api [None req-d857d01b-4657-47db-afda-873bd808eb35 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52f3cf4a-ab37-e791-e9aa-67bc02a3c5d2, 'name': SearchDatastore_Task, 'duration_secs': 0.015477} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.049610] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d857d01b-4657-47db-afda-873bd808eb35 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.056650] env[63202]: DEBUG oslo_vmware.api [None req-889ec7f6-4453-4d63-bd3b-38aa55b11694 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385625, 'name': CloneVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.072531] env[63202]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 959.072531] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52899f9a-b78c-208d-e9eb-7c975e3bc8d4" [ 959.072531] env[63202]: _type = "HttpNfcLease" [ 959.072531] env[63202]: } is ready. {{(pid=63202) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 959.072860] env[63202]: DEBUG oslo_vmware.rw_handles [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 959.072860] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52899f9a-b78c-208d-e9eb-7c975e3bc8d4" [ 959.072860] env[63202]: _type = "HttpNfcLease" [ 959.072860] env[63202]: }. {{(pid=63202) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 959.073616] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-343c6ea1-6e25-41cb-ae0e-95b2c4263ea7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.081625] env[63202]: DEBUG oslo_vmware.rw_handles [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52bc2395-25aa-f3fc-c935-b4787a11229d/disk-0.vmdk from lease info. {{(pid=63202) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 959.082123] env[63202]: DEBUG oslo_vmware.rw_handles [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52bc2395-25aa-f3fc-c935-b4787a11229d/disk-0.vmdk. {{(pid=63202) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 959.154013] env[63202]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-4d26f2d0-6d7c-49d6-9325-fc08e406c37a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.169405] env[63202]: ERROR nova.scheduler.client.report [None req-7de759cc-c39d-4058-8625-09664c11b490 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [req-b542d3bd-2f6a-4e5f-924d-952e879afa1d] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 79b33d17-3e75-494c-a550-67b275de2079. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-b542d3bd-2f6a-4e5f-924d-952e879afa1d"}]} [ 959.191922] env[63202]: DEBUG oslo_vmware.api [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385624, 'name': PowerOnVM_Task, 'duration_secs': 0.617901} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.193091] env[63202]: DEBUG nova.scheduler.client.report [None req-7de759cc-c39d-4058-8625-09664c11b490 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Refreshing inventories for resource provider 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 959.194829] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 959.195193] env[63202]: INFO nova.compute.manager [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Took 10.30 seconds to spawn the instance on the hypervisor. [ 959.195274] env[63202]: DEBUG nova.compute.manager [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 959.196252] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4bb6624-a2f9-45e2-b49a-00ec6611ea54 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.210576] env[63202]: DEBUG nova.scheduler.client.report [None req-7de759cc-c39d-4058-8625-09664c11b490 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Updating ProviderTree inventory for provider 79b33d17-3e75-494c-a550-67b275de2079 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 959.210758] env[63202]: DEBUG nova.compute.provider_tree [None req-7de759cc-c39d-4058-8625-09664c11b490 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Updating inventory in ProviderTree for provider 79b33d17-3e75-494c-a550-67b275de2079 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 959.226303] env[63202]: DEBUG nova.scheduler.client.report [None req-7de759cc-c39d-4058-8625-09664c11b490 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Refreshing aggregate associations for resource provider 79b33d17-3e75-494c-a550-67b275de2079, aggregates: None {{(pid=63202) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 959.244250] env[63202]: INFO nova.compute.manager [-] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Took 1.50 seconds to deallocate network for instance. [ 959.261093] env[63202]: DEBUG nova.scheduler.client.report [None req-7de759cc-c39d-4058-8625-09664c11b490 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Refreshing trait associations for resource provider 79b33d17-3e75-494c-a550-67b275de2079, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,HW_ARCH_X86_64,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO {{(pid=63202) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 959.296163] env[63202]: INFO nova.compute.manager [-] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Took 1.46 seconds to deallocate network for instance. [ 959.538853] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae7a7abc-fa7c-4c4b-8c48-ae4e4dd67d00 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.552364] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59b4c5e1-dd1b-41aa-8373-fd2426315e89 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.564506] env[63202]: DEBUG oslo_vmware.api [None req-889ec7f6-4453-4d63-bd3b-38aa55b11694 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385625, 'name': CloneVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.597388] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7782364-db13-4e07-be3f-103705b5ffe9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.605158] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67f07779-07b2-4a09-8e51-103e3a0e74ce {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.622934] env[63202]: DEBUG nova.compute.provider_tree [None req-7de759cc-c39d-4058-8625-09664c11b490 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Updating inventory in ProviderTree for provider 79b33d17-3e75-494c-a550-67b275de2079 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 959.629382] env[63202]: DEBUG nova.compute.manager [req-e4ce4872-4860-4f34-9e87-36e79a106b52 req-9af728ab-04fc-42d8-8f88-5ca775cec14a service nova] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Received event network-vif-deleted-718f088e-508c-4da3-97ec-11c84bfdab74 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 959.717101] env[63202]: INFO nova.compute.manager [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Took 28.91 seconds to build instance. [ 959.754250] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dcb77ba6-82bd-49fa-bf90-c76e177d7388 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.804062] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7e26c414-ad5f-4538-9c98-9fcc1de5bbcb tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.034374] env[63202]: DEBUG oslo_vmware.rw_handles [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Completed reading data from the image iterator. {{(pid=63202) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 960.034769] env[63202]: DEBUG oslo_vmware.rw_handles [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52bc2395-25aa-f3fc-c935-b4787a11229d/disk-0.vmdk. {{(pid=63202) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 960.036033] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef60d63c-cdc6-44b1-91b6-2e6fe8cbd385 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.044153] env[63202]: DEBUG oslo_vmware.rw_handles [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52bc2395-25aa-f3fc-c935-b4787a11229d/disk-0.vmdk is in state: ready. {{(pid=63202) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 960.044391] env[63202]: DEBUG oslo_vmware.rw_handles [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52bc2395-25aa-f3fc-c935-b4787a11229d/disk-0.vmdk. {{(pid=63202) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 960.044658] env[63202]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-2a8c2955-6c31-45c8-92c1-1dad11f4e47f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.058068] env[63202]: DEBUG oslo_vmware.api [None req-889ec7f6-4453-4d63-bd3b-38aa55b11694 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385625, 'name': CloneVM_Task} progress is 94%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.221458] env[63202]: DEBUG oslo_concurrency.lockutils [None req-70b450fb-c999-4d66-bfed-c9c6ef019a31 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "2110dfa7-1795-4c8f-92bf-18921409c99f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.575s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.245822] env[63202]: DEBUG nova.scheduler.client.report [None req-7de759cc-c39d-4058-8625-09664c11b490 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Updated inventory for provider 79b33d17-3e75-494c-a550-67b275de2079 with generation 97 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 960.246411] env[63202]: DEBUG nova.compute.provider_tree [None req-7de759cc-c39d-4058-8625-09664c11b490 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Updating resource provider 79b33d17-3e75-494c-a550-67b275de2079 generation from 97 to 98 during operation: update_inventory {{(pid=63202) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 960.246830] env[63202]: DEBUG nova.compute.provider_tree [None req-7de759cc-c39d-4058-8625-09664c11b490 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Updating inventory in ProviderTree for provider 79b33d17-3e75-494c-a550-67b275de2079 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 960.446738] env[63202]: DEBUG oslo_vmware.rw_handles [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52bc2395-25aa-f3fc-c935-b4787a11229d/disk-0.vmdk. {{(pid=63202) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 960.446738] env[63202]: INFO nova.virt.vmwareapi.images [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: da285417-bd38-4387-8521-df326ca0e326] Downloaded image file data 376aee6c-1f98-4a60-858b-40c55f65b90e [ 960.447404] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d53bfa3-3590-4ed0-9b52-40d819e04012 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.463599] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f4affe1b-be0c-4361-8160-64f8d3bd8b3b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.486360] env[63202]: INFO nova.virt.vmwareapi.images [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: da285417-bd38-4387-8521-df326ca0e326] The imported VM was unregistered [ 960.488602] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: da285417-bd38-4387-8521-df326ca0e326] Caching image {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 960.488844] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Creating directory with path [datastore1] devstack-image-cache_base/376aee6c-1f98-4a60-858b-40c55f65b90e {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 960.489142] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ddc40691-cb0b-4793-8f31-328497af1807 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.502538] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Created directory with path [datastore1] devstack-image-cache_base/376aee6c-1f98-4a60-858b-40c55f65b90e {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 960.503735] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_8b975109-76f2-4f78-a4b5-029f151227fc/OSTACK_IMG_8b975109-76f2-4f78-a4b5-029f151227fc.vmdk to [datastore1] devstack-image-cache_base/376aee6c-1f98-4a60-858b-40c55f65b90e/376aee6c-1f98-4a60-858b-40c55f65b90e.vmdk. {{(pid=63202) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 960.504047] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-bbd49ed0-019e-4e14-a752-518b958526d4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.510271] env[63202]: DEBUG oslo_vmware.api [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for the task: (returnval){ [ 960.510271] env[63202]: value = "task-1385627" [ 960.510271] env[63202]: _type = "Task" [ 960.510271] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.518615] env[63202]: DEBUG oslo_vmware.api [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385627, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.558609] env[63202]: DEBUG oslo_vmware.api [None req-889ec7f6-4453-4d63-bd3b-38aa55b11694 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385625, 'name': CloneVM_Task} progress is 94%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.760833] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7de759cc-c39d-4058-8625-09664c11b490 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.071s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.762065] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f52d6525-76d4-472b-8523-a728a64d8d38 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.773s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.762481] env[63202]: DEBUG nova.objects.instance [None req-f52d6525-76d4-472b-8523-a728a64d8d38 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Lazy-loading 'resources' on Instance uuid fc7d3924-a624-4ea4-890c-7628595fb733 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 960.788679] env[63202]: INFO nova.scheduler.client.report [None req-7de759cc-c39d-4058-8625-09664c11b490 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Deleted allocations for instance 3d1518d3-bed2-4b7b-af1d-d7da49e92874 [ 961.026240] env[63202]: DEBUG oslo_vmware.api [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385627, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.064085] env[63202]: DEBUG oslo_vmware.api [None req-889ec7f6-4453-4d63-bd3b-38aa55b11694 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385625, 'name': CloneVM_Task, 'duration_secs': 1.791058} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.064440] env[63202]: INFO nova.virt.vmwareapi.vmops [None req-889ec7f6-4453-4d63-bd3b-38aa55b11694 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Created linked-clone VM from snapshot [ 961.065550] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3caa68aa-4040-40c7-b564-40f0fb2411a9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.075423] env[63202]: DEBUG nova.virt.vmwareapi.images [None req-889ec7f6-4453-4d63-bd3b-38aa55b11694 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Uploading image b84c5f51-e408-48e5-a64c-0e8f168800d4 {{(pid=63202) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 961.089505] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-889ec7f6-4453-4d63-bd3b-38aa55b11694 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Destroying the VM {{(pid=63202) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 961.089807] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-922a8bcd-16ab-4095-9084-512507832516 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.098587] env[63202]: DEBUG oslo_vmware.api [None req-889ec7f6-4453-4d63-bd3b-38aa55b11694 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Waiting for the task: (returnval){ [ 961.098587] env[63202]: value = "task-1385628" [ 961.098587] env[63202]: _type = "Task" [ 961.098587] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.111609] env[63202]: DEBUG oslo_vmware.api [None req-889ec7f6-4453-4d63-bd3b-38aa55b11694 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385628, 'name': Destroy_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.306750] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7de759cc-c39d-4058-8625-09664c11b490 tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Lock "3d1518d3-bed2-4b7b-af1d-d7da49e92874" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.688s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.528342] env[63202]: DEBUG oslo_vmware.api [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385627, 'name': MoveVirtualDisk_Task} progress is 38%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.561432] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8000f19f-cf70-4e55-a2e1-5ee3b428f714 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.570939] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c530206-0f7d-4683-90f1-c6db7b9e58ef {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.617655] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e304209d-7257-4b95-902c-2aa3243b3bbe {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.631998] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5b5d27f-9f0a-4925-b4cf-6df4006eec69 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.637408] env[63202]: DEBUG oslo_vmware.api [None req-889ec7f6-4453-4d63-bd3b-38aa55b11694 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385628, 'name': Destroy_Task} progress is 33%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.650752] env[63202]: DEBUG nova.compute.provider_tree [None req-f52d6525-76d4-472b-8523-a728a64d8d38 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 961.673924] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c7c48f20-5023-47c4-9d97-1e62eb54e72e tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Acquiring lock "36291628-af7d-43cf-8149-09f57df47890" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.674286] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c7c48f20-5023-47c4-9d97-1e62eb54e72e tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Lock "36291628-af7d-43cf-8149-09f57df47890" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.674599] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c7c48f20-5023-47c4-9d97-1e62eb54e72e tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Acquiring lock "36291628-af7d-43cf-8149-09f57df47890-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.674813] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c7c48f20-5023-47c4-9d97-1e62eb54e72e tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Lock "36291628-af7d-43cf-8149-09f57df47890-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.674994] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c7c48f20-5023-47c4-9d97-1e62eb54e72e tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Lock "36291628-af7d-43cf-8149-09f57df47890-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.677511] env[63202]: INFO nova.compute.manager [None req-c7c48f20-5023-47c4-9d97-1e62eb54e72e tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] Terminating instance [ 961.679837] env[63202]: DEBUG nova.compute.manager [None req-c7c48f20-5023-47c4-9d97-1e62eb54e72e tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 961.680734] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c7c48f20-5023-47c4-9d97-1e62eb54e72e tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 961.681983] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56b4b941-0a90-472c-9c07-aaec01db61aa {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.690657] env[63202]: DEBUG oslo_concurrency.lockutils [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "eca51705-a972-48f6-85f5-6c397dad955d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.691228] env[63202]: DEBUG oslo_concurrency.lockutils [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "eca51705-a972-48f6-85f5-6c397dad955d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.695968] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7c48f20-5023-47c4-9d97-1e62eb54e72e tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 961.696613] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e0b2b316-567c-468f-af62-a1bb0db0bf73 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.704272] env[63202]: DEBUG oslo_vmware.api [None req-c7c48f20-5023-47c4-9d97-1e62eb54e72e tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Waiting for the task: (returnval){ [ 961.704272] env[63202]: value = "task-1385629" [ 961.704272] env[63202]: _type = "Task" [ 961.704272] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.715487] env[63202]: DEBUG oslo_vmware.api [None req-c7c48f20-5023-47c4-9d97-1e62eb54e72e tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385629, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.024781] env[63202]: DEBUG oslo_vmware.api [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385627, 'name': MoveVirtualDisk_Task} progress is 57%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.122335] env[63202]: DEBUG oslo_vmware.api [None req-889ec7f6-4453-4d63-bd3b-38aa55b11694 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385628, 'name': Destroy_Task, 'duration_secs': 0.582838} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.122709] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-889ec7f6-4453-4d63-bd3b-38aa55b11694 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Destroyed the VM [ 962.123083] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-889ec7f6-4453-4d63-bd3b-38aa55b11694 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Deleting Snapshot of the VM instance {{(pid=63202) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 962.123724] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-9908ab5c-79db-4257-8a66-ecd727196553 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.130929] env[63202]: DEBUG oslo_vmware.api [None req-889ec7f6-4453-4d63-bd3b-38aa55b11694 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Waiting for the task: (returnval){ [ 962.130929] env[63202]: value = "task-1385630" [ 962.130929] env[63202]: _type = "Task" [ 962.130929] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.139785] env[63202]: DEBUG oslo_vmware.api [None req-889ec7f6-4453-4d63-bd3b-38aa55b11694 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385630, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.154129] env[63202]: DEBUG nova.scheduler.client.report [None req-f52d6525-76d4-472b-8523-a728a64d8d38 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 962.197529] env[63202]: DEBUG nova.compute.manager [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 962.216104] env[63202]: DEBUG oslo_vmware.api [None req-c7c48f20-5023-47c4-9d97-1e62eb54e72e tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385629, 'name': PowerOffVM_Task, 'duration_secs': 0.317194} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.216427] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7c48f20-5023-47c4-9d97-1e62eb54e72e tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 962.216607] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c7c48f20-5023-47c4-9d97-1e62eb54e72e tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 962.216899] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3e385f6f-dd3a-4959-8700-744abafbb860 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.348073] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c7c48f20-5023-47c4-9d97-1e62eb54e72e tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 962.348073] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c7c48f20-5023-47c4-9d97-1e62eb54e72e tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] Deleting contents of the VM from datastore datastore2 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 962.348073] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7c48f20-5023-47c4-9d97-1e62eb54e72e tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Deleting the datastore file [datastore2] 36291628-af7d-43cf-8149-09f57df47890 {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 962.348073] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f0b743d1-5454-4d01-9d9b-54462595a043 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.356942] env[63202]: DEBUG oslo_vmware.api [None req-c7c48f20-5023-47c4-9d97-1e62eb54e72e tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Waiting for the task: (returnval){ [ 962.356942] env[63202]: value = "task-1385632" [ 962.356942] env[63202]: _type = "Task" [ 962.356942] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.366683] env[63202]: DEBUG oslo_vmware.api [None req-c7c48f20-5023-47c4-9d97-1e62eb54e72e tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385632, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.527690] env[63202]: DEBUG oslo_vmware.api [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385627, 'name': MoveVirtualDisk_Task} progress is 77%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.642155] env[63202]: DEBUG oslo_vmware.api [None req-889ec7f6-4453-4d63-bd3b-38aa55b11694 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385630, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.659685] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f52d6525-76d4-472b-8523-a728a64d8d38 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.897s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.662637] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 7.879s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.691412] env[63202]: INFO nova.scheduler.client.report [None req-f52d6525-76d4-472b-8523-a728a64d8d38 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Deleted allocations for instance fc7d3924-a624-4ea4-890c-7628595fb733 [ 962.722504] env[63202]: DEBUG oslo_concurrency.lockutils [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.866198] env[63202]: DEBUG oslo_vmware.api [None req-c7c48f20-5023-47c4-9d97-1e62eb54e72e tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Task: {'id': task-1385632, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.489589} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.866643] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7c48f20-5023-47c4-9d97-1e62eb54e72e tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 962.866766] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c7c48f20-5023-47c4-9d97-1e62eb54e72e tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] Deleted contents of the VM from datastore datastore2 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 962.866835] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c7c48f20-5023-47c4-9d97-1e62eb54e72e tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 962.867012] env[63202]: INFO nova.compute.manager [None req-c7c48f20-5023-47c4-9d97-1e62eb54e72e tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] [instance: 36291628-af7d-43cf-8149-09f57df47890] Took 1.19 seconds to destroy the instance on the hypervisor. [ 962.867285] env[63202]: DEBUG oslo.service.loopingcall [None req-c7c48f20-5023-47c4-9d97-1e62eb54e72e tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 962.867477] env[63202]: DEBUG nova.compute.manager [-] [instance: 36291628-af7d-43cf-8149-09f57df47890] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 962.867869] env[63202]: DEBUG nova.network.neutron [-] [instance: 36291628-af7d-43cf-8149-09f57df47890] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 963.029568] env[63202]: DEBUG oslo_vmware.api [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385627, 'name': MoveVirtualDisk_Task} progress is 97%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.146031] env[63202]: DEBUG oslo_vmware.api [None req-889ec7f6-4453-4d63-bd3b-38aa55b11694 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385630, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.208932] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f52d6525-76d4-472b-8523-a728a64d8d38 tempest-ServerShowV257Test-1709988600 tempest-ServerShowV257Test-1709988600-project-member] Lock "fc7d3924-a624-4ea4-890c-7628595fb733" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.958s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.228131] env[63202]: DEBUG nova.compute.manager [req-c2734201-a81d-4fae-8a27-fa6374e57cb6 req-ce87bcfd-c89b-4280-b025-59188ccffdd0 service nova] [instance: 36291628-af7d-43cf-8149-09f57df47890] Received event network-vif-deleted-355608b4-3a2c-4f49-8b9a-6b379a97934d {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 963.228398] env[63202]: INFO nova.compute.manager [req-c2734201-a81d-4fae-8a27-fa6374e57cb6 req-ce87bcfd-c89b-4280-b025-59188ccffdd0 service nova] [instance: 36291628-af7d-43cf-8149-09f57df47890] Neutron deleted interface 355608b4-3a2c-4f49-8b9a-6b379a97934d; detaching it from the instance and deleting it from the info cache [ 963.228598] env[63202]: DEBUG nova.network.neutron [req-c2734201-a81d-4fae-8a27-fa6374e57cb6 req-ce87bcfd-c89b-4280-b025-59188ccffdd0 service nova] [instance: 36291628-af7d-43cf-8149-09f57df47890] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.528013] env[63202]: DEBUG oslo_vmware.api [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385627, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.622206} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.528068] env[63202]: INFO nova.virt.vmwareapi.ds_util [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_8b975109-76f2-4f78-a4b5-029f151227fc/OSTACK_IMG_8b975109-76f2-4f78-a4b5-029f151227fc.vmdk to [datastore1] devstack-image-cache_base/376aee6c-1f98-4a60-858b-40c55f65b90e/376aee6c-1f98-4a60-858b-40c55f65b90e.vmdk. [ 963.528238] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: da285417-bd38-4387-8521-df326ca0e326] Cleaning up location [datastore1] OSTACK_IMG_8b975109-76f2-4f78-a4b5-029f151227fc {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 963.528439] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_8b975109-76f2-4f78-a4b5-029f151227fc {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 963.528698] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8af66902-88a4-41e8-9110-8cbb9af3ae4b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.534361] env[63202]: DEBUG oslo_vmware.api [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for the task: (returnval){ [ 963.534361] env[63202]: value = "task-1385633" [ 963.534361] env[63202]: _type = "Task" [ 963.534361] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.542277] env[63202]: DEBUG oslo_vmware.api [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385633, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.644609] env[63202]: DEBUG oslo_vmware.api [None req-889ec7f6-4453-4d63-bd3b-38aa55b11694 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385630, 'name': RemoveSnapshot_Task, 'duration_secs': 1.337241} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.645019] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-889ec7f6-4453-4d63-bd3b-38aa55b11694 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Deleted Snapshot of the VM instance {{(pid=63202) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 963.676720] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Applying migration context for instance e775e5e1-521a-4fc7-80e6-bcb6a70516c5 as it has an incoming, in-progress migration 78063a2a-a85c-4117-89c4-77c0fd76a7ae. Migration status is confirming {{(pid=63202) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 963.681072] env[63202]: INFO nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Updating resource usage from migration 78063a2a-a85c-4117-89c4-77c0fd76a7ae [ 963.697693] env[63202]: DEBUG nova.network.neutron [-] [instance: 36291628-af7d-43cf-8149-09f57df47890] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.703181] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 963.703315] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 36291628-af7d-43cf-8149-09f57df47890 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 963.704159] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 963.704159] env[63202]: WARNING nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance b1bccea9-2d79-431a-8be0-0a5ab293542a is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 963.704159] env[63202]: WARNING nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance f3652744-e072-4700-80d4-b9eca414c5cb is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 963.704159] env[63202]: WARNING nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 8755bceb-d510-4429-bd98-d6a63faf739d is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 963.704159] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance f3128c09-3680-4b0b-b463-3d6cd203fcf4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 963.704366] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 7437595c-fa35-483e-95f3-b75405b6bd13 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 963.704366] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Migration 78063a2a-a85c-4117-89c4-77c0fd76a7ae is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 963.704366] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance e775e5e1-521a-4fc7-80e6-bcb6a70516c5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 963.704366] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance a961b2a0-39c1-4267-9229-068e2b6ecd67 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 963.704476] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 2110dfa7-1795-4c8f-92bf-18921409c99f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 963.707518] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance da285417-bd38-4387-8521-df326ca0e326 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 963.731667] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9843aabb-a730-47dc-affb-c1829e77262a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.741431] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e27a85dd-e2ec-463c-ae78-bbd0f3454529 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.775090] env[63202]: DEBUG nova.compute.manager [req-c2734201-a81d-4fae-8a27-fa6374e57cb6 req-ce87bcfd-c89b-4280-b025-59188ccffdd0 service nova] [instance: 36291628-af7d-43cf-8149-09f57df47890] Detach interface failed, port_id=355608b4-3a2c-4f49-8b9a-6b379a97934d, reason: Instance 36291628-af7d-43cf-8149-09f57df47890 could not be found. {{(pid=63202) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 964.044917] env[63202]: DEBUG oslo_vmware.api [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385633, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.034537} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.047533] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 964.047533] env[63202]: DEBUG oslo_concurrency.lockutils [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Releasing lock "[datastore1] devstack-image-cache_base/376aee6c-1f98-4a60-858b-40c55f65b90e/376aee6c-1f98-4a60-858b-40c55f65b90e.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.047533] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/376aee6c-1f98-4a60-858b-40c55f65b90e/376aee6c-1f98-4a60-858b-40c55f65b90e.vmdk to [datastore1] da285417-bd38-4387-8521-df326ca0e326/da285417-bd38-4387-8521-df326ca0e326.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 964.047533] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c773f905-e570-4295-8537-81777201d887 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.053603] env[63202]: DEBUG oslo_vmware.api [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for the task: (returnval){ [ 964.053603] env[63202]: value = "task-1385637" [ 964.053603] env[63202]: _type = "Task" [ 964.053603] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.062467] env[63202]: DEBUG oslo_vmware.api [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385637, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.153182] env[63202]: WARNING nova.compute.manager [None req-889ec7f6-4453-4d63-bd3b-38aa55b11694 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Image not found during snapshot: nova.exception.ImageNotFound: Image b84c5f51-e408-48e5-a64c-0e8f168800d4 could not be found. [ 964.200054] env[63202]: INFO nova.compute.manager [-] [instance: 36291628-af7d-43cf-8149-09f57df47890] Took 1.33 seconds to deallocate network for instance. [ 964.207017] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 8437f856-a707-49c5-b8eb-5a22cdb990f8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 964.565033] env[63202]: DEBUG oslo_vmware.api [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385637, 'name': CopyVirtualDisk_Task} progress is 18%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.706794] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c7c48f20-5023-47c4-9d97-1e62eb54e72e tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.713841] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance eca51705-a972-48f6-85f5-6c397dad955d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 964.714140] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Total usable vcpus: 48, total allocated vcpus: 10 {{(pid=63202) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 964.714294] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2496MB phys_disk=200GB used_disk=10GB total_vcpus=48 used_vcpus=10 pci_stats=[] {{(pid=63202) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 964.850200] env[63202]: DEBUG oslo_concurrency.lockutils [None req-431db165-1743-4e2f-b49e-60e0686cf129 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Acquiring lock "a961b2a0-39c1-4267-9229-068e2b6ecd67" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.850540] env[63202]: DEBUG oslo_concurrency.lockutils [None req-431db165-1743-4e2f-b49e-60e0686cf129 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Lock "a961b2a0-39c1-4267-9229-068e2b6ecd67" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.851063] env[63202]: DEBUG oslo_concurrency.lockutils [None req-431db165-1743-4e2f-b49e-60e0686cf129 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Acquiring lock "a961b2a0-39c1-4267-9229-068e2b6ecd67-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.851319] env[63202]: DEBUG oslo_concurrency.lockutils [None req-431db165-1743-4e2f-b49e-60e0686cf129 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Lock "a961b2a0-39c1-4267-9229-068e2b6ecd67-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.851578] env[63202]: DEBUG oslo_concurrency.lockutils [None req-431db165-1743-4e2f-b49e-60e0686cf129 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Lock "a961b2a0-39c1-4267-9229-068e2b6ecd67-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.853777] env[63202]: INFO nova.compute.manager [None req-431db165-1743-4e2f-b49e-60e0686cf129 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Terminating instance [ 964.856293] env[63202]: DEBUG nova.compute.manager [None req-431db165-1743-4e2f-b49e-60e0686cf129 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 964.856699] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-431db165-1743-4e2f-b49e-60e0686cf129 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 964.858082] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea03b3d8-3a59-44c4-a058-75bf10ff42c4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.869441] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-431db165-1743-4e2f-b49e-60e0686cf129 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 964.869708] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-26646620-d1fd-4097-99ac-effe0fc85d4e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.876839] env[63202]: DEBUG oslo_vmware.api [None req-431db165-1743-4e2f-b49e-60e0686cf129 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Waiting for the task: (returnval){ [ 964.876839] env[63202]: value = "task-1385638" [ 964.876839] env[63202]: _type = "Task" [ 964.876839] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.887593] env[63202]: DEBUG oslo_vmware.api [None req-431db165-1743-4e2f-b49e-60e0686cf129 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385638, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.946439] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eaaa3ae-b024-4fa1-9048-601633314a9f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.954553] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcc7c735-b683-4e50-a150-5deac5adfbbb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.986646] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd9daa82-84a7-47b7-a394-628e83a98084 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.995093] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26765907-b60d-4653-85a5-78fcaa4b2942 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.008785] env[63202]: DEBUG nova.compute.provider_tree [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 965.063530] env[63202]: DEBUG oslo_vmware.api [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385637, 'name': CopyVirtualDisk_Task} progress is 38%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.386736] env[63202]: DEBUG oslo_vmware.api [None req-431db165-1743-4e2f-b49e-60e0686cf129 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385638, 'name': PowerOffVM_Task, 'duration_secs': 0.265918} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.387038] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-431db165-1743-4e2f-b49e-60e0686cf129 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 965.387220] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-431db165-1743-4e2f-b49e-60e0686cf129 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 965.387484] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-890c7795-cec6-4508-a82a-a252fb2fafdc {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.452858] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-431db165-1743-4e2f-b49e-60e0686cf129 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 965.453105] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-431db165-1743-4e2f-b49e-60e0686cf129 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Deleting contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 965.453475] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-431db165-1743-4e2f-b49e-60e0686cf129 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Deleting the datastore file [datastore1] a961b2a0-39c1-4267-9229-068e2b6ecd67 {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 965.453659] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-625bd2cb-5f19-4fe1-8ac0-7842d4bbf7bc {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.459470] env[63202]: DEBUG oslo_vmware.api [None req-431db165-1743-4e2f-b49e-60e0686cf129 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Waiting for the task: (returnval){ [ 965.459470] env[63202]: value = "task-1385640" [ 965.459470] env[63202]: _type = "Task" [ 965.459470] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.467921] env[63202]: DEBUG oslo_vmware.api [None req-431db165-1743-4e2f-b49e-60e0686cf129 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385640, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.513160] env[63202]: DEBUG nova.scheduler.client.report [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 965.564651] env[63202]: DEBUG oslo_vmware.api [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385637, 'name': CopyVirtualDisk_Task} progress is 60%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.972913] env[63202]: DEBUG oslo_vmware.api [None req-431db165-1743-4e2f-b49e-60e0686cf129 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385640, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.017954] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63202) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 966.018242] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.356s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.018545] env[63202]: DEBUG oslo_concurrency.lockutils [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.325s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.020468] env[63202]: INFO nova.compute.claims [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 966.065599] env[63202]: DEBUG oslo_vmware.api [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385637, 'name': CopyVirtualDisk_Task} progress is 80%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.395440] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 966.395757] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Starting heal instance info cache {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 966.471060] env[63202]: DEBUG oslo_vmware.api [None req-431db165-1743-4e2f-b49e-60e0686cf129 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385640, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.568333] env[63202]: DEBUG oslo_vmware.api [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385637, 'name': CopyVirtualDisk_Task} progress is 94%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.972046] env[63202]: DEBUG oslo_vmware.api [None req-431db165-1743-4e2f-b49e-60e0686cf129 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385640, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.351171} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.972354] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-431db165-1743-4e2f-b49e-60e0686cf129 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 966.972546] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-431db165-1743-4e2f-b49e-60e0686cf129 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Deleted contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 966.972733] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-431db165-1743-4e2f-b49e-60e0686cf129 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 966.972950] env[63202]: INFO nova.compute.manager [None req-431db165-1743-4e2f-b49e-60e0686cf129 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Took 2.12 seconds to destroy the instance on the hypervisor. [ 966.973257] env[63202]: DEBUG oslo.service.loopingcall [None req-431db165-1743-4e2f-b49e-60e0686cf129 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 966.973472] env[63202]: DEBUG nova.compute.manager [-] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 966.974124] env[63202]: DEBUG nova.network.neutron [-] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 967.069196] env[63202]: DEBUG oslo_vmware.api [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385637, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.683469} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.070434] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/376aee6c-1f98-4a60-858b-40c55f65b90e/376aee6c-1f98-4a60-858b-40c55f65b90e.vmdk to [datastore1] da285417-bd38-4387-8521-df326ca0e326/da285417-bd38-4387-8521-df326ca0e326.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 967.072142] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3a6aa70-b958-4b59-8d4f-44237308a66c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.096930] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: da285417-bd38-4387-8521-df326ca0e326] Reconfiguring VM instance instance-00000054 to attach disk [datastore1] da285417-bd38-4387-8521-df326ca0e326/da285417-bd38-4387-8521-df326ca0e326.vmdk or device None with type streamOptimized {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 967.100970] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-97c1e11c-fa13-4d63-8ae4-226690247b1e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.124023] env[63202]: DEBUG oslo_vmware.api [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for the task: (returnval){ [ 967.124023] env[63202]: value = "task-1385642" [ 967.124023] env[63202]: _type = "Task" [ 967.124023] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.131889] env[63202]: DEBUG oslo_vmware.api [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385642, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.287383] env[63202]: DEBUG nova.compute.manager [req-58810c7a-4093-40a6-abe9-1ce5d21db294 req-7277428a-0398-4804-a72e-5dad7cd17cd4 service nova] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Received event network-vif-deleted-836478e8-9299-44c3-a4b0-957cd1706387 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 967.287690] env[63202]: INFO nova.compute.manager [req-58810c7a-4093-40a6-abe9-1ce5d21db294 req-7277428a-0398-4804-a72e-5dad7cd17cd4 service nova] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Neutron deleted interface 836478e8-9299-44c3-a4b0-957cd1706387; detaching it from the instance and deleting it from the info cache [ 967.287951] env[63202]: DEBUG nova.network.neutron [req-58810c7a-4093-40a6-abe9-1ce5d21db294 req-7277428a-0398-4804-a72e-5dad7cd17cd4 service nova] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.299726] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c78cfcb-f2df-4434-9fe3-e20e22f6eb3a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.309532] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e4cd4de-0cb3-400e-b9a7-2fb0e7bf9aa6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.348019] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edff58e9-28e9-428f-8101-023bec02588a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.355309] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bedd0c3-80d2-4943-b55f-835f1e03f462 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.370156] env[63202]: DEBUG nova.compute.provider_tree [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 967.632150] env[63202]: DEBUG oslo_vmware.api [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385642, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.754280] env[63202]: DEBUG nova.network.neutron [-] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.790635] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f8abf125-12df-4e7c-8ef1-7a3b39346d16 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.801651] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fa0415d-fe84-4609-b585-587104b07b61 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.829733] env[63202]: DEBUG nova.compute.manager [req-58810c7a-4093-40a6-abe9-1ce5d21db294 req-7277428a-0398-4804-a72e-5dad7cd17cd4 service nova] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Detach interface failed, port_id=836478e8-9299-44c3-a4b0-957cd1706387, reason: Instance a961b2a0-39c1-4267-9229-068e2b6ecd67 could not be found. {{(pid=63202) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 967.873611] env[63202]: DEBUG nova.scheduler.client.report [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 968.134116] env[63202]: DEBUG oslo_vmware.api [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385642, 'name': ReconfigVM_Task, 'duration_secs': 0.523003} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.134468] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: da285417-bd38-4387-8521-df326ca0e326] Reconfigured VM instance instance-00000054 to attach disk [datastore1] da285417-bd38-4387-8521-df326ca0e326/da285417-bd38-4387-8521-df326ca0e326.vmdk or device None with type streamOptimized {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 968.135038] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5f3dd817-6d43-4899-9802-16d8ecdffc52 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.141022] env[63202]: DEBUG oslo_vmware.api [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for the task: (returnval){ [ 968.141022] env[63202]: value = "task-1385643" [ 968.141022] env[63202]: _type = "Task" [ 968.141022] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.149042] env[63202]: DEBUG oslo_vmware.api [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385643, 'name': Rename_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.257509] env[63202]: INFO nova.compute.manager [-] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Took 1.28 seconds to deallocate network for instance. [ 968.378850] env[63202]: DEBUG oslo_concurrency.lockutils [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.360s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.379415] env[63202]: DEBUG nova.compute.manager [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 968.382174] env[63202]: DEBUG oslo_concurrency.lockutils [None req-854b8b38-28af-4f84-ac47-238da88ee0d4 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.472s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.382371] env[63202]: DEBUG oslo_concurrency.lockutils [None req-854b8b38-28af-4f84-ac47-238da88ee0d4 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.384739] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d857d01b-4657-47db-afda-873bd808eb35 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 9.335s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.405100] env[63202]: INFO nova.scheduler.client.report [None req-854b8b38-28af-4f84-ac47-238da88ee0d4 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Deleted allocations for instance 8755bceb-d510-4429-bd98-d6a63faf739d [ 968.423567] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Acquiring lock "refresh_cache-36291628-af7d-43cf-8149-09f57df47890" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.423567] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Acquired lock "refresh_cache-36291628-af7d-43cf-8149-09f57df47890" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.423708] env[63202]: DEBUG nova.network.neutron [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 36291628-af7d-43cf-8149-09f57df47890] Forcefully refreshing network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 968.653731] env[63202]: DEBUG oslo_vmware.api [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385643, 'name': Rename_Task} progress is 99%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.764058] env[63202]: DEBUG oslo_concurrency.lockutils [None req-431db165-1743-4e2f-b49e-60e0686cf129 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.888551] env[63202]: DEBUG nova.compute.utils [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 968.893292] env[63202]: DEBUG nova.compute.manager [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 968.893771] env[63202]: DEBUG nova.network.neutron [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 968.915378] env[63202]: DEBUG oslo_concurrency.lockutils [None req-854b8b38-28af-4f84-ac47-238da88ee0d4 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "8755bceb-d510-4429-bd98-d6a63faf739d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.868s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.950147] env[63202]: DEBUG nova.network.neutron [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 36291628-af7d-43cf-8149-09f57df47890] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 968.962215] env[63202]: DEBUG nova.policy [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3e9b38cd253441939b048166110aef47', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2b2de35030a9484094e964ffc30a822d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 969.103624] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48d88980-5b47-437e-a5ca-e8940084f8d9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.111546] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-452a35ac-3d6f-4e4c-bbed-493fec74f559 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.146034] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2e13e1e-eb04-481d-b3d4-cd823efa73d3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.155308] env[63202]: DEBUG oslo_vmware.api [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385643, 'name': Rename_Task} progress is 99%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.156558] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5209245b-8881-472f-b945-5ea3d672202e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.169550] env[63202]: DEBUG nova.compute.provider_tree [None req-d857d01b-4657-47db-afda-873bd808eb35 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 969.396178] env[63202]: DEBUG nova.compute.manager [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 969.446438] env[63202]: DEBUG nova.network.neutron [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Successfully created port: 07cc827a-876a-4432-a5fc-9ba8920dc5f9 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 969.558675] env[63202]: DEBUG nova.network.neutron [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 36291628-af7d-43cf-8149-09f57df47890] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.651332] env[63202]: DEBUG oslo_vmware.api [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385643, 'name': Rename_Task, 'duration_secs': 1.130927} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.651796] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: da285417-bd38-4387-8521-df326ca0e326] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 969.652096] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-443e11c4-479c-4256-9fb6-9740fc8c118c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.657976] env[63202]: DEBUG oslo_vmware.api [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for the task: (returnval){ [ 969.657976] env[63202]: value = "task-1385645" [ 969.657976] env[63202]: _type = "Task" [ 969.657976] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.667179] env[63202]: DEBUG oslo_vmware.api [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385645, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.672460] env[63202]: DEBUG nova.scheduler.client.report [None req-d857d01b-4657-47db-afda-873bd808eb35 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 970.062223] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Releasing lock "refresh_cache-36291628-af7d-43cf-8149-09f57df47890" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.062435] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 36291628-af7d-43cf-8149-09f57df47890] Updated the network info_cache for instance {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 970.062639] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 970.062925] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 970.168109] env[63202]: DEBUG oslo_vmware.api [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385645, 'name': PowerOnVM_Task, 'duration_secs': 0.456302} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.168421] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: da285417-bd38-4387-8521-df326ca0e326] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 970.168669] env[63202]: INFO nova.compute.manager [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: da285417-bd38-4387-8521-df326ca0e326] Took 17.33 seconds to spawn the instance on the hypervisor. [ 970.168826] env[63202]: DEBUG nova.compute.manager [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: da285417-bd38-4387-8521-df326ca0e326] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 970.169577] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e7af8ce-44c8-409f-ab5d-b978676b007f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.407585] env[63202]: DEBUG nova.compute.manager [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 970.433215] env[63202]: DEBUG nova.virt.hardware [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 970.433709] env[63202]: DEBUG nova.virt.hardware [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 970.433888] env[63202]: DEBUG nova.virt.hardware [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 970.434100] env[63202]: DEBUG nova.virt.hardware [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 970.434270] env[63202]: DEBUG nova.virt.hardware [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 970.434407] env[63202]: DEBUG nova.virt.hardware [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 970.434636] env[63202]: DEBUG nova.virt.hardware [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 970.434796] env[63202]: DEBUG nova.virt.hardware [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 970.434967] env[63202]: DEBUG nova.virt.hardware [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 970.435144] env[63202]: DEBUG nova.virt.hardware [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 970.435321] env[63202]: DEBUG nova.virt.hardware [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 970.436204] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23078f4e-732c-4f5b-9dc6-b17df8ccbf71 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.444493] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-767a12bc-57e6-46f0-b2fb-4e815f78a06d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.605465] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "1bfb520e-6708-45a0-bb1f-2037fe1d801e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.605750] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "1bfb520e-6708-45a0-bb1f-2037fe1d801e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.686024] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d857d01b-4657-47db-afda-873bd808eb35 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.301s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.690295] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dcb77ba6-82bd-49fa-bf90-c76e177d7388 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.936s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.690553] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dcb77ba6-82bd-49fa-bf90-c76e177d7388 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.692478] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7e26c414-ad5f-4538-9c98-9fcc1de5bbcb tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.889s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.692672] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7e26c414-ad5f-4538-9c98-9fcc1de5bbcb tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.694305] env[63202]: DEBUG oslo_concurrency.lockutils [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.973s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.695622] env[63202]: INFO nova.compute.claims [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 970.698037] env[63202]: INFO nova.compute.manager [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: da285417-bd38-4387-8521-df326ca0e326] Took 33.77 seconds to build instance. [ 970.724602] env[63202]: INFO nova.scheduler.client.report [None req-7e26c414-ad5f-4538-9c98-9fcc1de5bbcb tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Deleted allocations for instance f3652744-e072-4700-80d4-b9eca414c5cb [ 970.726477] env[63202]: INFO nova.scheduler.client.report [None req-dcb77ba6-82bd-49fa-bf90-c76e177d7388 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Deleted allocations for instance b1bccea9-2d79-431a-8be0-0a5ab293542a [ 970.963605] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f2b62a87-5eb1-4d78-93c5-10354cf0ea0e tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquiring lock "da285417-bd38-4387-8521-df326ca0e326" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.108395] env[63202]: DEBUG nova.compute.manager [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 971.179720] env[63202]: DEBUG nova.compute.manager [req-db70e3dd-04cf-48ae-b6ca-e03c7d3cc3c4 req-3750fe78-1f89-497a-b72d-a2f15e8ee8cc service nova] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Received event network-vif-plugged-07cc827a-876a-4432-a5fc-9ba8920dc5f9 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 971.180277] env[63202]: DEBUG oslo_concurrency.lockutils [req-db70e3dd-04cf-48ae-b6ca-e03c7d3cc3c4 req-3750fe78-1f89-497a-b72d-a2f15e8ee8cc service nova] Acquiring lock "8437f856-a707-49c5-b8eb-5a22cdb990f8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.180550] env[63202]: DEBUG oslo_concurrency.lockutils [req-db70e3dd-04cf-48ae-b6ca-e03c7d3cc3c4 req-3750fe78-1f89-497a-b72d-a2f15e8ee8cc service nova] Lock "8437f856-a707-49c5-b8eb-5a22cdb990f8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.180737] env[63202]: DEBUG oslo_concurrency.lockutils [req-db70e3dd-04cf-48ae-b6ca-e03c7d3cc3c4 req-3750fe78-1f89-497a-b72d-a2f15e8ee8cc service nova] Lock "8437f856-a707-49c5-b8eb-5a22cdb990f8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.180955] env[63202]: DEBUG nova.compute.manager [req-db70e3dd-04cf-48ae-b6ca-e03c7d3cc3c4 req-3750fe78-1f89-497a-b72d-a2f15e8ee8cc service nova] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] No waiting events found dispatching network-vif-plugged-07cc827a-876a-4432-a5fc-9ba8920dc5f9 {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 971.181164] env[63202]: WARNING nova.compute.manager [req-db70e3dd-04cf-48ae-b6ca-e03c7d3cc3c4 req-3750fe78-1f89-497a-b72d-a2f15e8ee8cc service nova] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Received unexpected event network-vif-plugged-07cc827a-876a-4432-a5fc-9ba8920dc5f9 for instance with vm_state building and task_state spawning. [ 971.207265] env[63202]: DEBUG oslo_concurrency.lockutils [None req-744e18b0-9e99-420c-b4b7-ad8c1dbb4d12 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lock "da285417-bd38-4387-8521-df326ca0e326" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.288s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.211115] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f2b62a87-5eb1-4d78-93c5-10354cf0ea0e tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lock "da285417-bd38-4387-8521-df326ca0e326" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.246s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.211115] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f2b62a87-5eb1-4d78-93c5-10354cf0ea0e tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquiring lock "da285417-bd38-4387-8521-df326ca0e326-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.211115] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f2b62a87-5eb1-4d78-93c5-10354cf0ea0e tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lock "da285417-bd38-4387-8521-df326ca0e326-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.211115] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f2b62a87-5eb1-4d78-93c5-10354cf0ea0e tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lock "da285417-bd38-4387-8521-df326ca0e326-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.212080] env[63202]: INFO nova.compute.manager [None req-f2b62a87-5eb1-4d78-93c5-10354cf0ea0e tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: da285417-bd38-4387-8521-df326ca0e326] Terminating instance [ 971.213870] env[63202]: DEBUG nova.compute.manager [None req-f2b62a87-5eb1-4d78-93c5-10354cf0ea0e tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: da285417-bd38-4387-8521-df326ca0e326] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 971.214097] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-f2b62a87-5eb1-4d78-93c5-10354cf0ea0e tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: da285417-bd38-4387-8521-df326ca0e326] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 971.215586] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8564ec53-605a-43a1-8859-fa845b64c7af {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.224484] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2b62a87-5eb1-4d78-93c5-10354cf0ea0e tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: da285417-bd38-4387-8521-df326ca0e326] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 971.224775] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c94917d1-e340-4080-bbe7-7646488dc4e5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.236500] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7e26c414-ad5f-4538-9c98-9fcc1de5bbcb tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Lock "f3652744-e072-4700-80d4-b9eca414c5cb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.561s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.239641] env[63202]: DEBUG oslo_vmware.api [None req-f2b62a87-5eb1-4d78-93c5-10354cf0ea0e tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for the task: (returnval){ [ 971.239641] env[63202]: value = "task-1385647" [ 971.239641] env[63202]: _type = "Task" [ 971.239641] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.240680] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dcb77ba6-82bd-49fa-bf90-c76e177d7388 tempest-ListServersNegativeTestJSON-1642626060 tempest-ListServersNegativeTestJSON-1642626060-project-member] Lock "b1bccea9-2d79-431a-8be0-0a5ab293542a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.656s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.250564] env[63202]: DEBUG oslo_vmware.api [None req-f2b62a87-5eb1-4d78-93c5-10354cf0ea0e tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385647, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.282218] env[63202]: INFO nova.scheduler.client.report [None req-d857d01b-4657-47db-afda-873bd808eb35 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Deleted allocation for migration 78063a2a-a85c-4117-89c4-77c0fd76a7ae [ 971.429986] env[63202]: DEBUG nova.network.neutron [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Successfully updated port: 07cc827a-876a-4432-a5fc-9ba8920dc5f9 {{(pid=63202) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 971.632068] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.750087] env[63202]: DEBUG oslo_vmware.api [None req-f2b62a87-5eb1-4d78-93c5-10354cf0ea0e tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385647, 'name': PowerOffVM_Task, 'duration_secs': 0.189195} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.752373] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2b62a87-5eb1-4d78-93c5-10354cf0ea0e tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: da285417-bd38-4387-8521-df326ca0e326] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 971.752833] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-f2b62a87-5eb1-4d78-93c5-10354cf0ea0e tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: da285417-bd38-4387-8521-df326ca0e326] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 971.753622] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ccf9b624-8f1f-45ae-96dd-b6b5ca3c7c30 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.787941] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d857d01b-4657-47db-afda-873bd808eb35 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "e775e5e1-521a-4fc7-80e6-bcb6a70516c5" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 15.952s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.825108] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-f2b62a87-5eb1-4d78-93c5-10354cf0ea0e tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: da285417-bd38-4387-8521-df326ca0e326] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 971.825360] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-f2b62a87-5eb1-4d78-93c5-10354cf0ea0e tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: da285417-bd38-4387-8521-df326ca0e326] Deleting contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 971.825545] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2b62a87-5eb1-4d78-93c5-10354cf0ea0e tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Deleting the datastore file [datastore1] da285417-bd38-4387-8521-df326ca0e326 {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 971.829572] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9e6133a3-0f68-45d4-b1cf-3c85798b6c8a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.837621] env[63202]: DEBUG oslo_vmware.api [None req-f2b62a87-5eb1-4d78-93c5-10354cf0ea0e tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for the task: (returnval){ [ 971.837621] env[63202]: value = "task-1385649" [ 971.837621] env[63202]: _type = "Task" [ 971.837621] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.847464] env[63202]: DEBUG oslo_vmware.api [None req-f2b62a87-5eb1-4d78-93c5-10354cf0ea0e tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385649, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.882023] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b23f32f-21dc-4212-9351-255909270a3d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.888209] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ca8a635-8261-47d7-866c-33683508fa19 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.918557] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3793481-09d1-42dc-a08e-2de776adee72 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.926387] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61001416-c56d-428a-aafa-525062ba334e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.940221] env[63202]: DEBUG oslo_concurrency.lockutils [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquiring lock "refresh_cache-8437f856-a707-49c5-b8eb-5a22cdb990f8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 971.940368] env[63202]: DEBUG oslo_concurrency.lockutils [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquired lock "refresh_cache-8437f856-a707-49c5-b8eb-5a22cdb990f8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.940649] env[63202]: DEBUG nova.network.neutron [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 971.942283] env[63202]: DEBUG nova.compute.provider_tree [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 972.347951] env[63202]: DEBUG oslo_vmware.api [None req-f2b62a87-5eb1-4d78-93c5-10354cf0ea0e tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385649, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.164161} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.348258] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2b62a87-5eb1-4d78-93c5-10354cf0ea0e tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 972.348258] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-f2b62a87-5eb1-4d78-93c5-10354cf0ea0e tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: da285417-bd38-4387-8521-df326ca0e326] Deleted contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 972.348464] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-f2b62a87-5eb1-4d78-93c5-10354cf0ea0e tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: da285417-bd38-4387-8521-df326ca0e326] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 972.349571] env[63202]: INFO nova.compute.manager [None req-f2b62a87-5eb1-4d78-93c5-10354cf0ea0e tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: da285417-bd38-4387-8521-df326ca0e326] Took 1.13 seconds to destroy the instance on the hypervisor. [ 972.349571] env[63202]: DEBUG oslo.service.loopingcall [None req-f2b62a87-5eb1-4d78-93c5-10354cf0ea0e tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 972.349571] env[63202]: DEBUG nova.compute.manager [-] [instance: da285417-bd38-4387-8521-df326ca0e326] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 972.349571] env[63202]: DEBUG nova.network.neutron [-] [instance: da285417-bd38-4387-8521-df326ca0e326] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 972.447223] env[63202]: DEBUG nova.scheduler.client.report [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 972.480631] env[63202]: DEBUG nova.network.neutron [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 972.723599] env[63202]: DEBUG nova.network.neutron [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Updating instance_info_cache with network_info: [{"id": "07cc827a-876a-4432-a5fc-9ba8920dc5f9", "address": "fa:16:3e:71:e6:9f", "network": {"id": "06ab5813-9ad9-4021-9bdb-f2f02af8d73f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1714653503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b2de35030a9484094e964ffc30a822d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3093647a-bab7-4562-ada0-428725e8c0fc", "external-id": "nsx-vlan-transportzone-660", "segmentation_id": 660, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap07cc827a-87", "ovs_interfaceid": "07cc827a-876a-4432-a5fc-9ba8920dc5f9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 972.952668] env[63202]: DEBUG oslo_concurrency.lockutils [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.258s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.952929] env[63202]: DEBUG nova.compute.manager [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 972.957719] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c7c48f20-5023-47c4-9d97-1e62eb54e72e tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.249s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.957719] env[63202]: DEBUG nova.objects.instance [None req-c7c48f20-5023-47c4-9d97-1e62eb54e72e tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Lazy-loading 'resources' on Instance uuid 36291628-af7d-43cf-8149-09f57df47890 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 973.211119] env[63202]: DEBUG nova.network.neutron [-] [instance: da285417-bd38-4387-8521-df326ca0e326] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.216030] env[63202]: DEBUG nova.compute.manager [req-b5758412-273c-4e4e-bb7d-641c78fd4b02 req-0c5741bb-a4a3-45ea-8829-cafacfef163b service nova] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Received event network-changed-07cc827a-876a-4432-a5fc-9ba8920dc5f9 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 973.216142] env[63202]: DEBUG nova.compute.manager [req-b5758412-273c-4e4e-bb7d-641c78fd4b02 req-0c5741bb-a4a3-45ea-8829-cafacfef163b service nova] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Refreshing instance network info cache due to event network-changed-07cc827a-876a-4432-a5fc-9ba8920dc5f9. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 973.216405] env[63202]: DEBUG oslo_concurrency.lockutils [req-b5758412-273c-4e4e-bb7d-641c78fd4b02 req-0c5741bb-a4a3-45ea-8829-cafacfef163b service nova] Acquiring lock "refresh_cache-8437f856-a707-49c5-b8eb-5a22cdb990f8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 973.225884] env[63202]: DEBUG oslo_concurrency.lockutils [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Releasing lock "refresh_cache-8437f856-a707-49c5-b8eb-5a22cdb990f8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.226334] env[63202]: DEBUG nova.compute.manager [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Instance network_info: |[{"id": "07cc827a-876a-4432-a5fc-9ba8920dc5f9", "address": "fa:16:3e:71:e6:9f", "network": {"id": "06ab5813-9ad9-4021-9bdb-f2f02af8d73f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1714653503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b2de35030a9484094e964ffc30a822d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3093647a-bab7-4562-ada0-428725e8c0fc", "external-id": "nsx-vlan-transportzone-660", "segmentation_id": 660, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap07cc827a-87", "ovs_interfaceid": "07cc827a-876a-4432-a5fc-9ba8920dc5f9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 973.230726] env[63202]: DEBUG oslo_concurrency.lockutils [req-b5758412-273c-4e4e-bb7d-641c78fd4b02 req-0c5741bb-a4a3-45ea-8829-cafacfef163b service nova] Acquired lock "refresh_cache-8437f856-a707-49c5-b8eb-5a22cdb990f8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.231110] env[63202]: DEBUG nova.network.neutron [req-b5758412-273c-4e4e-bb7d-641c78fd4b02 req-0c5741bb-a4a3-45ea-8829-cafacfef163b service nova] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Refreshing network info cache for port 07cc827a-876a-4432-a5fc-9ba8920dc5f9 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 973.233141] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:71:e6:9f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3093647a-bab7-4562-ada0-428725e8c0fc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '07cc827a-876a-4432-a5fc-9ba8920dc5f9', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 973.247875] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Creating folder: Project (2b2de35030a9484094e964ffc30a822d). Parent ref: group-v294090. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 973.247875] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9c038c47-3cf8-42ab-b677-34156c552d90 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.258073] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Created folder: Project (2b2de35030a9484094e964ffc30a822d) in parent group-v294090. [ 973.258348] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Creating folder: Instances. Parent ref: group-v294183. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 973.258642] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1f889e69-5f9b-4fc7-8cd1-3c5051185828 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.270038] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Created folder: Instances in parent group-v294183. [ 973.270038] env[63202]: DEBUG oslo.service.loopingcall [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 973.270038] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 973.270038] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-89f40a2c-defb-4209-8817-7f66d75cbe6f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.296950] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 973.296950] env[63202]: value = "task-1385653" [ 973.296950] env[63202]: _type = "Task" [ 973.296950] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.304912] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385653, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.459283] env[63202]: DEBUG nova.compute.utils [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 973.464861] env[63202]: DEBUG nova.compute.manager [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 973.465073] env[63202]: DEBUG nova.network.neutron [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: eca51705-a972-48f6-85f5-6c397dad955d] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 973.516403] env[63202]: DEBUG nova.policy [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2bd91973e1bb4ce7be8bce9bfe34a941', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e972faeaedc6468aab7e7cfee88a477b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 973.666620] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbd9d7ec-0802-4a34-abc8-6d2069057c40 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.674979] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ca0de4b-815d-434c-984c-af5d7a6264bf {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.711447] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c279ffe9-e7b1-4303-bbc4-ac60d3fea105 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.719338] env[63202]: INFO nova.compute.manager [-] [instance: da285417-bd38-4387-8521-df326ca0e326] Took 1.37 seconds to deallocate network for instance. [ 973.722175] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-becf7c18-118a-4f43-895c-0333e32de1da {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.740502] env[63202]: DEBUG nova.compute.provider_tree [None req-c7c48f20-5023-47c4-9d97-1e62eb54e72e tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 973.802128] env[63202]: DEBUG nova.network.neutron [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Successfully created port: 145b6d77-5716-400a-911a-37decf12fae3 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 973.809354] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385653, 'name': CreateVM_Task, 'duration_secs': 0.309736} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.809769] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 973.810498] env[63202]: DEBUG oslo_concurrency.lockutils [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 973.810680] env[63202]: DEBUG oslo_concurrency.lockutils [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.811070] env[63202]: DEBUG oslo_concurrency.lockutils [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 973.811355] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-48a93bce-0320-485e-9b22-a25e06737cb9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.815700] env[63202]: DEBUG oslo_vmware.api [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 973.815700] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52fa1d3b-cb92-44c3-94d7-c7fa29241d2e" [ 973.815700] env[63202]: _type = "Task" [ 973.815700] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.824154] env[63202]: DEBUG oslo_vmware.api [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52fa1d3b-cb92-44c3-94d7-c7fa29241d2e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.966448] env[63202]: DEBUG nova.compute.manager [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 974.083061] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e6c16b26-535e-43d0-9def-02a7399b5a99 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "e775e5e1-521a-4fc7-80e6-bcb6a70516c5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.083061] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e6c16b26-535e-43d0-9def-02a7399b5a99 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "e775e5e1-521a-4fc7-80e6-bcb6a70516c5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.083061] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e6c16b26-535e-43d0-9def-02a7399b5a99 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "e775e5e1-521a-4fc7-80e6-bcb6a70516c5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.083061] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e6c16b26-535e-43d0-9def-02a7399b5a99 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "e775e5e1-521a-4fc7-80e6-bcb6a70516c5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.083423] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e6c16b26-535e-43d0-9def-02a7399b5a99 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "e775e5e1-521a-4fc7-80e6-bcb6a70516c5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.085608] env[63202]: INFO nova.compute.manager [None req-e6c16b26-535e-43d0-9def-02a7399b5a99 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Terminating instance [ 974.088255] env[63202]: DEBUG nova.compute.manager [None req-e6c16b26-535e-43d0-9def-02a7399b5a99 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 974.088545] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-e6c16b26-535e-43d0-9def-02a7399b5a99 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 974.089630] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77e6d11b-2b7c-4ddb-9955-c7f6902b5f79 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.098175] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6c16b26-535e-43d0-9def-02a7399b5a99 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 974.098547] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5b3c65b3-3850-4628-a977-6eeded428ff5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.105198] env[63202]: DEBUG oslo_vmware.api [None req-e6c16b26-535e-43d0-9def-02a7399b5a99 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for the task: (returnval){ [ 974.105198] env[63202]: value = "task-1385654" [ 974.105198] env[63202]: _type = "Task" [ 974.105198] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.115322] env[63202]: DEBUG oslo_vmware.api [None req-e6c16b26-535e-43d0-9def-02a7399b5a99 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385654, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.167387] env[63202]: DEBUG nova.network.neutron [req-b5758412-273c-4e4e-bb7d-641c78fd4b02 req-0c5741bb-a4a3-45ea-8829-cafacfef163b service nova] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Updated VIF entry in instance network info cache for port 07cc827a-876a-4432-a5fc-9ba8920dc5f9. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 974.167781] env[63202]: DEBUG nova.network.neutron [req-b5758412-273c-4e4e-bb7d-641c78fd4b02 req-0c5741bb-a4a3-45ea-8829-cafacfef163b service nova] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Updating instance_info_cache with network_info: [{"id": "07cc827a-876a-4432-a5fc-9ba8920dc5f9", "address": "fa:16:3e:71:e6:9f", "network": {"id": "06ab5813-9ad9-4021-9bdb-f2f02af8d73f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1714653503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b2de35030a9484094e964ffc30a822d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3093647a-bab7-4562-ada0-428725e8c0fc", "external-id": "nsx-vlan-transportzone-660", "segmentation_id": 660, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap07cc827a-87", "ovs_interfaceid": "07cc827a-876a-4432-a5fc-9ba8920dc5f9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.232256] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f2b62a87-5eb1-4d78-93c5-10354cf0ea0e tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.243634] env[63202]: DEBUG nova.scheduler.client.report [None req-c7c48f20-5023-47c4-9d97-1e62eb54e72e tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 974.327883] env[63202]: DEBUG oslo_vmware.api [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52fa1d3b-cb92-44c3-94d7-c7fa29241d2e, 'name': SearchDatastore_Task, 'duration_secs': 0.011133} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.328245] env[63202]: DEBUG oslo_concurrency.lockutils [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.328497] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 974.328764] env[63202]: DEBUG oslo_concurrency.lockutils [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.328918] env[63202]: DEBUG oslo_concurrency.lockutils [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.329132] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 974.329437] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-77c949d6-0088-4f6c-a5f1-679e8074bae0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.337855] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 974.338068] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 974.338984] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8d2a5d79-0c71-4169-b663-50456f9cd3da {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.346500] env[63202]: DEBUG oslo_vmware.api [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 974.346500] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]524c6293-b1f1-4ffd-1bc4-6f74c048cae6" [ 974.346500] env[63202]: _type = "Task" [ 974.346500] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.354302] env[63202]: DEBUG oslo_vmware.api [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]524c6293-b1f1-4ffd-1bc4-6f74c048cae6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.619174] env[63202]: DEBUG oslo_vmware.api [None req-e6c16b26-535e-43d0-9def-02a7399b5a99 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385654, 'name': PowerOffVM_Task, 'duration_secs': 0.199602} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.619453] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6c16b26-535e-43d0-9def-02a7399b5a99 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 974.619631] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-e6c16b26-535e-43d0-9def-02a7399b5a99 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 974.619855] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-24da7b89-eb92-429a-a1d7-ebd92c694a0a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.667959] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "dd146c71-c391-41e8-8cc7-7276f4a518d9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.668277] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "dd146c71-c391-41e8-8cc7-7276f4a518d9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.670568] env[63202]: DEBUG oslo_concurrency.lockutils [req-b5758412-273c-4e4e-bb7d-641c78fd4b02 req-0c5741bb-a4a3-45ea-8829-cafacfef163b service nova] Releasing lock "refresh_cache-8437f856-a707-49c5-b8eb-5a22cdb990f8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.670822] env[63202]: DEBUG nova.compute.manager [req-b5758412-273c-4e4e-bb7d-641c78fd4b02 req-0c5741bb-a4a3-45ea-8829-cafacfef163b service nova] [instance: da285417-bd38-4387-8521-df326ca0e326] Received event network-vif-deleted-a0547ee8-b263-49d2-8152-cbac29b8c25b {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 974.671012] env[63202]: INFO nova.compute.manager [req-b5758412-273c-4e4e-bb7d-641c78fd4b02 req-0c5741bb-a4a3-45ea-8829-cafacfef163b service nova] [instance: da285417-bd38-4387-8521-df326ca0e326] Neutron deleted interface a0547ee8-b263-49d2-8152-cbac29b8c25b; detaching it from the instance and deleting it from the info cache [ 974.671369] env[63202]: DEBUG nova.network.neutron [req-b5758412-273c-4e4e-bb7d-641c78fd4b02 req-0c5741bb-a4a3-45ea-8829-cafacfef163b service nova] [instance: da285417-bd38-4387-8521-df326ca0e326] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.749449] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c7c48f20-5023-47c4-9d97-1e62eb54e72e tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.794s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.752489] env[63202]: DEBUG oslo_concurrency.lockutils [None req-431db165-1743-4e2f-b49e-60e0686cf129 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.988s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.752785] env[63202]: DEBUG nova.objects.instance [None req-431db165-1743-4e2f-b49e-60e0686cf129 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Lazy-loading 'resources' on Instance uuid a961b2a0-39c1-4267-9229-068e2b6ecd67 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 974.774134] env[63202]: INFO nova.scheduler.client.report [None req-c7c48f20-5023-47c4-9d97-1e62eb54e72e tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Deleted allocations for instance 36291628-af7d-43cf-8149-09f57df47890 [ 974.857725] env[63202]: DEBUG oslo_vmware.api [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]524c6293-b1f1-4ffd-1bc4-6f74c048cae6, 'name': SearchDatastore_Task, 'duration_secs': 0.008279} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.858665] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fcad81be-b0c0-4d70-8349-fef28d8d49b6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.864070] env[63202]: DEBUG oslo_vmware.api [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 974.864070] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52aac316-786b-4516-c9b3-38a02afc7028" [ 974.864070] env[63202]: _type = "Task" [ 974.864070] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.872354] env[63202]: DEBUG oslo_vmware.api [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52aac316-786b-4516-c9b3-38a02afc7028, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.897666] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-e6c16b26-535e-43d0-9def-02a7399b5a99 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 974.897894] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-e6c16b26-535e-43d0-9def-02a7399b5a99 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Deleting contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 974.898092] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6c16b26-535e-43d0-9def-02a7399b5a99 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Deleting the datastore file [datastore1] e775e5e1-521a-4fc7-80e6-bcb6a70516c5 {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 974.898620] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-01a0a09c-d0f3-48a1-abe2-7c8e2a3e75e6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.904854] env[63202]: DEBUG oslo_vmware.api [None req-e6c16b26-535e-43d0-9def-02a7399b5a99 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for the task: (returnval){ [ 974.904854] env[63202]: value = "task-1385656" [ 974.904854] env[63202]: _type = "Task" [ 974.904854] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.912545] env[63202]: DEBUG oslo_vmware.api [None req-e6c16b26-535e-43d0-9def-02a7399b5a99 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385656, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.978553] env[63202]: DEBUG nova.compute.manager [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 975.008933] env[63202]: DEBUG nova.virt.hardware [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 975.009392] env[63202]: DEBUG nova.virt.hardware [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 975.009671] env[63202]: DEBUG nova.virt.hardware [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 975.009997] env[63202]: DEBUG nova.virt.hardware [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 975.010252] env[63202]: DEBUG nova.virt.hardware [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 975.010493] env[63202]: DEBUG nova.virt.hardware [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 975.010784] env[63202]: DEBUG nova.virt.hardware [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 975.010966] env[63202]: DEBUG nova.virt.hardware [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 975.011244] env[63202]: DEBUG nova.virt.hardware [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 975.011549] env[63202]: DEBUG nova.virt.hardware [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 975.011786] env[63202]: DEBUG nova.virt.hardware [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 975.012701] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77178fdc-b006-4778-a8e6-1e455baaf09d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.021058] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-105da448-f73a-4728-b2eb-652af21856a0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.171314] env[63202]: DEBUG nova.compute.manager [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 975.175580] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c6e866dc-e659-431e-ba48-c5e07a8942ef {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.184985] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60f3f15c-aa26-46f4-88d7-acb998aa3930 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.214631] env[63202]: DEBUG nova.compute.manager [req-b5758412-273c-4e4e-bb7d-641c78fd4b02 req-0c5741bb-a4a3-45ea-8829-cafacfef163b service nova] [instance: da285417-bd38-4387-8521-df326ca0e326] Detach interface failed, port_id=a0547ee8-b263-49d2-8152-cbac29b8c25b, reason: Instance da285417-bd38-4387-8521-df326ca0e326 could not be found. {{(pid=63202) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 975.236876] env[63202]: DEBUG nova.compute.manager [req-9c1ba4a9-ff2b-40d1-b85b-49f31aa6b9a4 req-329c3ec5-6e99-41ec-a2fe-2bb72ecce62c service nova] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Received event network-vif-plugged-145b6d77-5716-400a-911a-37decf12fae3 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 975.237122] env[63202]: DEBUG oslo_concurrency.lockutils [req-9c1ba4a9-ff2b-40d1-b85b-49f31aa6b9a4 req-329c3ec5-6e99-41ec-a2fe-2bb72ecce62c service nova] Acquiring lock "eca51705-a972-48f6-85f5-6c397dad955d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.237390] env[63202]: DEBUG oslo_concurrency.lockutils [req-9c1ba4a9-ff2b-40d1-b85b-49f31aa6b9a4 req-329c3ec5-6e99-41ec-a2fe-2bb72ecce62c service nova] Lock "eca51705-a972-48f6-85f5-6c397dad955d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.237496] env[63202]: DEBUG oslo_concurrency.lockutils [req-9c1ba4a9-ff2b-40d1-b85b-49f31aa6b9a4 req-329c3ec5-6e99-41ec-a2fe-2bb72ecce62c service nova] Lock "eca51705-a972-48f6-85f5-6c397dad955d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.237669] env[63202]: DEBUG nova.compute.manager [req-9c1ba4a9-ff2b-40d1-b85b-49f31aa6b9a4 req-329c3ec5-6e99-41ec-a2fe-2bb72ecce62c service nova] [instance: eca51705-a972-48f6-85f5-6c397dad955d] No waiting events found dispatching network-vif-plugged-145b6d77-5716-400a-911a-37decf12fae3 {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 975.237852] env[63202]: WARNING nova.compute.manager [req-9c1ba4a9-ff2b-40d1-b85b-49f31aa6b9a4 req-329c3ec5-6e99-41ec-a2fe-2bb72ecce62c service nova] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Received unexpected event network-vif-plugged-145b6d77-5716-400a-911a-37decf12fae3 for instance with vm_state building and task_state spawning. [ 975.282711] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c7c48f20-5023-47c4-9d97-1e62eb54e72e tempest-ListServerFiltersTestJSON-974110797 tempest-ListServerFiltersTestJSON-974110797-project-member] Lock "36291628-af7d-43cf-8149-09f57df47890" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.608s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.345058] env[63202]: DEBUG nova.network.neutron [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Successfully updated port: 145b6d77-5716-400a-911a-37decf12fae3 {{(pid=63202) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 975.379321] env[63202]: DEBUG oslo_vmware.api [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52aac316-786b-4516-c9b3-38a02afc7028, 'name': SearchDatastore_Task, 'duration_secs': 0.014898} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.379605] env[63202]: DEBUG oslo_concurrency.lockutils [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.379872] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore2] 8437f856-a707-49c5-b8eb-5a22cdb990f8/8437f856-a707-49c5-b8eb-5a22cdb990f8.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 975.380162] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d943b858-4473-40f6-9dff-cf4a6fda6a0d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.388878] env[63202]: DEBUG oslo_vmware.api [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 975.388878] env[63202]: value = "task-1385658" [ 975.388878] env[63202]: _type = "Task" [ 975.388878] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.401171] env[63202]: DEBUG oslo_vmware.api [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385658, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.416879] env[63202]: DEBUG oslo_vmware.api [None req-e6c16b26-535e-43d0-9def-02a7399b5a99 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385656, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.279517} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.417203] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6c16b26-535e-43d0-9def-02a7399b5a99 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 975.417463] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-e6c16b26-535e-43d0-9def-02a7399b5a99 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Deleted contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 975.417688] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-e6c16b26-535e-43d0-9def-02a7399b5a99 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 975.417936] env[63202]: INFO nova.compute.manager [None req-e6c16b26-535e-43d0-9def-02a7399b5a99 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Took 1.33 seconds to destroy the instance on the hypervisor. [ 975.418250] env[63202]: DEBUG oslo.service.loopingcall [None req-e6c16b26-535e-43d0-9def-02a7399b5a99 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 975.418531] env[63202]: DEBUG nova.compute.manager [-] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 975.418677] env[63202]: DEBUG nova.network.neutron [-] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 975.422210] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3d36189-a6a8-4f59-b1ce-61d4458b65eb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.431257] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f23a0ae0-f991-4577-a005-82813a4e2a64 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.464689] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-617f6bcf-cdfd-4bc5-89fc-0854a470d858 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.472759] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-178372d2-8231-4910-8115-53883cae9daa {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.486318] env[63202]: DEBUG nova.compute.provider_tree [None req-431db165-1743-4e2f-b49e-60e0686cf129 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 975.689995] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.849341] env[63202]: DEBUG oslo_concurrency.lockutils [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "refresh_cache-eca51705-a972-48f6-85f5-6c397dad955d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.849513] env[63202]: DEBUG oslo_concurrency.lockutils [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquired lock "refresh_cache-eca51705-a972-48f6-85f5-6c397dad955d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.849799] env[63202]: DEBUG nova.network.neutron [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 975.902728] env[63202]: DEBUG oslo_vmware.api [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385658, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.991872] env[63202]: DEBUG nova.scheduler.client.report [None req-431db165-1743-4e2f-b49e-60e0686cf129 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 976.161161] env[63202]: DEBUG nova.network.neutron [-] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 976.389836] env[63202]: DEBUG nova.network.neutron [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 976.400787] env[63202]: DEBUG oslo_vmware.api [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385658, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.667031} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.401094] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore2] 8437f856-a707-49c5-b8eb-5a22cdb990f8/8437f856-a707-49c5-b8eb-5a22cdb990f8.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 976.401338] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 976.401625] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-268862c9-2870-4e8f-b044-cabe3b3f972d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.408163] env[63202]: DEBUG oslo_vmware.api [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 976.408163] env[63202]: value = "task-1385659" [ 976.408163] env[63202]: _type = "Task" [ 976.408163] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.417046] env[63202]: DEBUG oslo_vmware.api [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385659, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.497925] env[63202]: DEBUG oslo_concurrency.lockutils [None req-431db165-1743-4e2f-b49e-60e0686cf129 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.745s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.500267] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.868s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.502324] env[63202]: INFO nova.compute.claims [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 976.526419] env[63202]: INFO nova.scheduler.client.report [None req-431db165-1743-4e2f-b49e-60e0686cf129 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Deleted allocations for instance a961b2a0-39c1-4267-9229-068e2b6ecd67 [ 976.550442] env[63202]: DEBUG nova.network.neutron [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Updating instance_info_cache with network_info: [{"id": "145b6d77-5716-400a-911a-37decf12fae3", "address": "fa:16:3e:04:01:35", "network": {"id": "b1ea2320-3801-43fe-b566-4f69db81329e", "bridge": "br-int", "label": "tempest-ServersTestJSON-618582355-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e972faeaedc6468aab7e7cfee88a477b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4223acd2-30f7-440e-b975-60b30d931694", "external-id": "nsx-vlan-transportzone-647", "segmentation_id": 647, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap145b6d77-57", "ovs_interfaceid": "145b6d77-5716-400a-911a-37decf12fae3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 976.663978] env[63202]: INFO nova.compute.manager [-] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Took 1.25 seconds to deallocate network for instance. [ 976.918715] env[63202]: DEBUG oslo_vmware.api [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385659, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065927} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.919017] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 976.919749] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc56ea10-fa87-4af8-8288-c261ee9e4344 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.942164] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] 8437f856-a707-49c5-b8eb-5a22cdb990f8/8437f856-a707-49c5-b8eb-5a22cdb990f8.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 976.942473] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-473e0835-465d-4a05-91b0-4152c1cbb291 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.960835] env[63202]: DEBUG oslo_vmware.api [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 976.960835] env[63202]: value = "task-1385660" [ 976.960835] env[63202]: _type = "Task" [ 976.960835] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.968142] env[63202]: DEBUG oslo_vmware.api [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385660, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.034348] env[63202]: DEBUG oslo_concurrency.lockutils [None req-431db165-1743-4e2f-b49e-60e0686cf129 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Lock "a961b2a0-39c1-4267-9229-068e2b6ecd67" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.184s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.053901] env[63202]: DEBUG oslo_concurrency.lockutils [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Releasing lock "refresh_cache-eca51705-a972-48f6-85f5-6c397dad955d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.054253] env[63202]: DEBUG nova.compute.manager [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Instance network_info: |[{"id": "145b6d77-5716-400a-911a-37decf12fae3", "address": "fa:16:3e:04:01:35", "network": {"id": "b1ea2320-3801-43fe-b566-4f69db81329e", "bridge": "br-int", "label": "tempest-ServersTestJSON-618582355-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e972faeaedc6468aab7e7cfee88a477b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4223acd2-30f7-440e-b975-60b30d931694", "external-id": "nsx-vlan-transportzone-647", "segmentation_id": 647, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap145b6d77-57", "ovs_interfaceid": "145b6d77-5716-400a-911a-37decf12fae3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 977.054951] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:04:01:35', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4223acd2-30f7-440e-b975-60b30d931694', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '145b6d77-5716-400a-911a-37decf12fae3', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 977.064258] env[63202]: DEBUG oslo.service.loopingcall [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 977.064802] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 977.065113] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ac83cab0-f2a6-4c2c-9656-3b0f1741dd9b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.084461] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 977.084461] env[63202]: value = "task-1385661" [ 977.084461] env[63202]: _type = "Task" [ 977.084461] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.092213] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385661, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.171090] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e6c16b26-535e-43d0-9def-02a7399b5a99 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.273488] env[63202]: DEBUG nova.compute.manager [req-93ae91a4-f2f5-4cfc-b1c2-3ee1a6fdbd68 req-55db0c9b-d58e-43ab-a5bc-c654fbe10ac3 service nova] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Received event network-changed-145b6d77-5716-400a-911a-37decf12fae3 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 977.273639] env[63202]: DEBUG nova.compute.manager [req-93ae91a4-f2f5-4cfc-b1c2-3ee1a6fdbd68 req-55db0c9b-d58e-43ab-a5bc-c654fbe10ac3 service nova] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Refreshing instance network info cache due to event network-changed-145b6d77-5716-400a-911a-37decf12fae3. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 977.273897] env[63202]: DEBUG oslo_concurrency.lockutils [req-93ae91a4-f2f5-4cfc-b1c2-3ee1a6fdbd68 req-55db0c9b-d58e-43ab-a5bc-c654fbe10ac3 service nova] Acquiring lock "refresh_cache-eca51705-a972-48f6-85f5-6c397dad955d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.274061] env[63202]: DEBUG oslo_concurrency.lockutils [req-93ae91a4-f2f5-4cfc-b1c2-3ee1a6fdbd68 req-55db0c9b-d58e-43ab-a5bc-c654fbe10ac3 service nova] Acquired lock "refresh_cache-eca51705-a972-48f6-85f5-6c397dad955d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.274236] env[63202]: DEBUG nova.network.neutron [req-93ae91a4-f2f5-4cfc-b1c2-3ee1a6fdbd68 req-55db0c9b-d58e-43ab-a5bc-c654fbe10ac3 service nova] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Refreshing network info cache for port 145b6d77-5716-400a-911a-37decf12fae3 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 977.470655] env[63202]: DEBUG oslo_vmware.api [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385660, 'name': ReconfigVM_Task, 'duration_secs': 0.331552} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.470939] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Reconfigured VM instance instance-00000055 to attach disk [datastore2] 8437f856-a707-49c5-b8eb-5a22cdb990f8/8437f856-a707-49c5-b8eb-5a22cdb990f8.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 977.471609] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ffcff4fd-ec93-4be2-b55e-81505056bca2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.477604] env[63202]: DEBUG oslo_vmware.api [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 977.477604] env[63202]: value = "task-1385663" [ 977.477604] env[63202]: _type = "Task" [ 977.477604] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.484950] env[63202]: DEBUG oslo_vmware.api [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385663, 'name': Rename_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.594499] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385661, 'name': CreateVM_Task, 'duration_secs': 0.372569} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.597149] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 977.598130] env[63202]: DEBUG oslo_concurrency.lockutils [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.598216] env[63202]: DEBUG oslo_concurrency.lockutils [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.598539] env[63202]: DEBUG oslo_concurrency.lockutils [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 977.598834] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-40e5452e-e782-482a-80ac-53f9fedacab3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.603484] env[63202]: DEBUG oslo_vmware.api [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 977.603484] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]5283a645-d622-9724-0eaa-090ebdce7208" [ 977.603484] env[63202]: _type = "Task" [ 977.603484] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.616361] env[63202]: DEBUG oslo_vmware.api [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]5283a645-d622-9724-0eaa-090ebdce7208, 'name': SearchDatastore_Task, 'duration_secs': 0.008591} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.617524] env[63202]: DEBUG oslo_concurrency.lockutils [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.617524] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 977.617524] env[63202]: DEBUG oslo_concurrency.lockutils [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.617524] env[63202]: DEBUG oslo_concurrency.lockutils [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.617524] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 977.617833] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5ff7e4da-76de-4d35-afd6-228ff872d034 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.625170] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 977.625238] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 977.626595] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-455411c3-a0c8-4ec4-867c-6c071529426f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.631607] env[63202]: DEBUG oslo_vmware.api [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 977.631607] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52aa6f40-1f1d-be13-a044-7aee5d3205f0" [ 977.631607] env[63202]: _type = "Task" [ 977.631607] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.643030] env[63202]: DEBUG oslo_vmware.api [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52aa6f40-1f1d-be13-a044-7aee5d3205f0, 'name': SearchDatastore_Task, 'duration_secs': 0.007315} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.645975] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a6619ed4-7f14-4873-b5a3-06e505a5ec26 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.650669] env[63202]: DEBUG oslo_vmware.api [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 977.650669] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52c9cc26-fe86-8e28-a401-aa4275d2f5a5" [ 977.650669] env[63202]: _type = "Task" [ 977.650669] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.660581] env[63202]: DEBUG oslo_vmware.api [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52c9cc26-fe86-8e28-a401-aa4275d2f5a5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.667782] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08a7c888-5033-492e-bc17-3f784c6f2ea3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.674265] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9297641f-6bcd-4f65-b0c8-fc8f6cbcb119 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.705153] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-230205ef-2c75-4ce0-be84-5da72d8dc929 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.712455] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2084eebf-8470-4777-9bdc-85333e654a39 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.725469] env[63202]: DEBUG nova.compute.provider_tree [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 977.986964] env[63202]: DEBUG oslo_vmware.api [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385663, 'name': Rename_Task, 'duration_secs': 0.135383} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.987285] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 977.987397] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e7b13197-1d44-454b-b67e-c0e3bafc9480 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.993308] env[63202]: DEBUG oslo_vmware.api [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 977.993308] env[63202]: value = "task-1385664" [ 977.993308] env[63202]: _type = "Task" [ 977.993308] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.996694] env[63202]: DEBUG nova.network.neutron [req-93ae91a4-f2f5-4cfc-b1c2-3ee1a6fdbd68 req-55db0c9b-d58e-43ab-a5bc-c654fbe10ac3 service nova] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Updated VIF entry in instance network info cache for port 145b6d77-5716-400a-911a-37decf12fae3. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 977.997064] env[63202]: DEBUG nova.network.neutron [req-93ae91a4-f2f5-4cfc-b1c2-3ee1a6fdbd68 req-55db0c9b-d58e-43ab-a5bc-c654fbe10ac3 service nova] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Updating instance_info_cache with network_info: [{"id": "145b6d77-5716-400a-911a-37decf12fae3", "address": "fa:16:3e:04:01:35", "network": {"id": "b1ea2320-3801-43fe-b566-4f69db81329e", "bridge": "br-int", "label": "tempest-ServersTestJSON-618582355-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e972faeaedc6468aab7e7cfee88a477b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4223acd2-30f7-440e-b975-60b30d931694", "external-id": "nsx-vlan-transportzone-647", "segmentation_id": 647, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap145b6d77-57", "ovs_interfaceid": "145b6d77-5716-400a-911a-37decf12fae3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.002879] env[63202]: DEBUG oslo_vmware.api [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385664, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.167553] env[63202]: DEBUG oslo_vmware.api [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52c9cc26-fe86-8e28-a401-aa4275d2f5a5, 'name': SearchDatastore_Task, 'duration_secs': 0.008952} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.167849] env[63202]: DEBUG oslo_concurrency.lockutils [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.168139] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] eca51705-a972-48f6-85f5-6c397dad955d/eca51705-a972-48f6-85f5-6c397dad955d.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 978.168466] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-004023d8-728d-465d-99ca-26c1ce46b2a9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.177915] env[63202]: DEBUG oslo_vmware.api [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 978.177915] env[63202]: value = "task-1385665" [ 978.177915] env[63202]: _type = "Task" [ 978.177915] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.189469] env[63202]: DEBUG oslo_vmware.api [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385665, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.229538] env[63202]: DEBUG nova.scheduler.client.report [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 978.500047] env[63202]: DEBUG oslo_concurrency.lockutils [req-93ae91a4-f2f5-4cfc-b1c2-3ee1a6fdbd68 req-55db0c9b-d58e-43ab-a5bc-c654fbe10ac3 service nova] Releasing lock "refresh_cache-eca51705-a972-48f6-85f5-6c397dad955d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.500282] env[63202]: DEBUG nova.compute.manager [req-93ae91a4-f2f5-4cfc-b1c2-3ee1a6fdbd68 req-55db0c9b-d58e-43ab-a5bc-c654fbe10ac3 service nova] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Received event network-vif-deleted-c7da6929-1bd1-4e2b-bcb8-609863cf6861 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 978.506379] env[63202]: DEBUG oslo_vmware.api [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385664, 'name': PowerOnVM_Task, 'duration_secs': 0.435546} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.506650] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 978.506858] env[63202]: INFO nova.compute.manager [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Took 8.10 seconds to spawn the instance on the hypervisor. [ 978.507070] env[63202]: DEBUG nova.compute.manager [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 978.507972] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68d708fa-82d5-4556-804c-cc98df391503 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.688785] env[63202]: DEBUG oslo_vmware.api [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385665, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.735509] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.235s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.736212] env[63202]: DEBUG nova.compute.manager [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 978.738897] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f2b62a87-5eb1-4d78-93c5-10354cf0ea0e tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.507s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.739140] env[63202]: DEBUG nova.objects.instance [None req-f2b62a87-5eb1-4d78-93c5-10354cf0ea0e tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lazy-loading 'resources' on Instance uuid da285417-bd38-4387-8521-df326ca0e326 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 979.024413] env[63202]: INFO nova.compute.manager [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Took 20.35 seconds to build instance. [ 979.063676] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a456b4ef-c33d-4853-a71d-3eab0fce779d tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Acquiring lock "0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.063943] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a456b4ef-c33d-4853-a71d-3eab0fce779d tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lock "0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.188742] env[63202]: DEBUG oslo_vmware.api [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385665, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.653519} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.188742] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] eca51705-a972-48f6-85f5-6c397dad955d/eca51705-a972-48f6-85f5-6c397dad955d.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 979.188876] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 979.189160] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-394907d1-5286-4ab3-a943-254489f28439 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.195545] env[63202]: DEBUG oslo_vmware.api [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 979.195545] env[63202]: value = "task-1385666" [ 979.195545] env[63202]: _type = "Task" [ 979.195545] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.205072] env[63202]: DEBUG oslo_vmware.api [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385666, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.242991] env[63202]: DEBUG nova.compute.utils [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 979.247534] env[63202]: DEBUG nova.compute.manager [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 979.247801] env[63202]: DEBUG nova.network.neutron [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 979.255850] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Acquiring lock "8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.255850] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Lock "8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.313587] env[63202]: DEBUG nova.policy [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b59e7e62d92e4f5eb0fee41e8f5196d7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f508ec04edc844a19640a8a85f27e5b9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 979.434828] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd45016b-974c-4392-b8f5-99e34d679a5c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.443220] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a628cb49-d52a-4518-8209-5d63aad566fa {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.480330] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5236c82d-e4fa-4209-919a-d35b1e948ba0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.488098] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bdd32ab-0e64-4a88-950d-95095e068358 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.501329] env[63202]: DEBUG nova.compute.provider_tree [None req-f2b62a87-5eb1-4d78-93c5-10354cf0ea0e tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 979.526780] env[63202]: DEBUG oslo_concurrency.lockutils [None req-37150549-40b4-4821-b512-fe56ac79c786 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lock "8437f856-a707-49c5-b8eb-5a22cdb990f8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.862s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.571101] env[63202]: INFO nova.compute.manager [None req-a456b4ef-c33d-4853-a71d-3eab0fce779d tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Detaching volume 06e1ab70-2c2e-4848-9e08-4e17242e47b5 [ 979.597750] env[63202]: DEBUG nova.network.neutron [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Successfully created port: df7c236b-8be0-46d0-9dc1-56cba5b73ec7 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 979.625172] env[63202]: INFO nova.virt.block_device [None req-a456b4ef-c33d-4853-a71d-3eab0fce779d tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Attempting to driver detach volume 06e1ab70-2c2e-4848-9e08-4e17242e47b5 from mountpoint /dev/sdb [ 979.626476] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-a456b4ef-c33d-4853-a71d-3eab0fce779d tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Volume detach. Driver type: vmdk {{(pid=63202) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 979.626668] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-a456b4ef-c33d-4853-a71d-3eab0fce779d tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294155', 'volume_id': '06e1ab70-2c2e-4848-9e08-4e17242e47b5', 'name': 'volume-06e1ab70-2c2e-4848-9e08-4e17242e47b5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb', 'attached_at': '', 'detached_at': '', 'volume_id': '06e1ab70-2c2e-4848-9e08-4e17242e47b5', 'serial': '06e1ab70-2c2e-4848-9e08-4e17242e47b5'} {{(pid=63202) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 979.627706] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d868c08-f279-4cc7-8653-50483666420b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.651159] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c85852c0-a8af-4e2d-a85f-a80cd4dcb8c0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.659829] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af20a57c-ea9b-4818-8edb-e49e1cc63747 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.685830] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3f0f8ae-09bd-4691-a90e-fe6d1bc8f8db {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.702660] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-a456b4ef-c33d-4853-a71d-3eab0fce779d tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] The volume has not been displaced from its original location: [datastore2] volume-06e1ab70-2c2e-4848-9e08-4e17242e47b5/volume-06e1ab70-2c2e-4848-9e08-4e17242e47b5.vmdk. No consolidation needed. {{(pid=63202) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 979.712358] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-a456b4ef-c33d-4853-a71d-3eab0fce779d tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Reconfiguring VM instance instance-0000003f to detach disk 2001 {{(pid=63202) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 979.712358] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-33601de7-4d11-4a6d-9452-3fcc9c86113f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.732255] env[63202]: DEBUG oslo_vmware.api [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385666, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070146} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.734628] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 979.734628] env[63202]: DEBUG oslo_vmware.api [None req-a456b4ef-c33d-4853-a71d-3eab0fce779d tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Waiting for the task: (returnval){ [ 979.734628] env[63202]: value = "task-1385668" [ 979.734628] env[63202]: _type = "Task" [ 979.734628] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.734907] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c2ea454-fbc5-44a2-866b-a0569fbec656 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.745444] env[63202]: DEBUG oslo_vmware.api [None req-a456b4ef-c33d-4853-a71d-3eab0fce779d tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385668, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.755052] env[63202]: DEBUG nova.compute.manager [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 979.766641] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Reconfiguring VM instance instance-00000056 to attach disk [datastore1] eca51705-a972-48f6-85f5-6c397dad955d/eca51705-a972-48f6-85f5-6c397dad955d.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 979.767368] env[63202]: DEBUG nova.compute.manager [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 979.769993] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d57791c2-ecb2-40b3-ab36-92295952c232 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.792438] env[63202]: DEBUG oslo_vmware.api [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 979.792438] env[63202]: value = "task-1385669" [ 979.792438] env[63202]: _type = "Task" [ 979.792438] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.802204] env[63202]: DEBUG oslo_vmware.api [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385669, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.004667] env[63202]: DEBUG nova.scheduler.client.report [None req-f2b62a87-5eb1-4d78-93c5-10354cf0ea0e tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 980.040433] env[63202]: DEBUG nova.compute.manager [req-7912c24b-66b7-45a9-b387-24c004ab62d3 req-1e8c7953-6023-48bc-acd0-b7495fcc620a service nova] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Received event network-changed-07cc827a-876a-4432-a5fc-9ba8920dc5f9 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 980.041028] env[63202]: DEBUG nova.compute.manager [req-7912c24b-66b7-45a9-b387-24c004ab62d3 req-1e8c7953-6023-48bc-acd0-b7495fcc620a service nova] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Refreshing instance network info cache due to event network-changed-07cc827a-876a-4432-a5fc-9ba8920dc5f9. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 980.041112] env[63202]: DEBUG oslo_concurrency.lockutils [req-7912c24b-66b7-45a9-b387-24c004ab62d3 req-1e8c7953-6023-48bc-acd0-b7495fcc620a service nova] Acquiring lock "refresh_cache-8437f856-a707-49c5-b8eb-5a22cdb990f8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.041318] env[63202]: DEBUG oslo_concurrency.lockutils [req-7912c24b-66b7-45a9-b387-24c004ab62d3 req-1e8c7953-6023-48bc-acd0-b7495fcc620a service nova] Acquired lock "refresh_cache-8437f856-a707-49c5-b8eb-5a22cdb990f8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.041595] env[63202]: DEBUG nova.network.neutron [req-7912c24b-66b7-45a9-b387-24c004ab62d3 req-1e8c7953-6023-48bc-acd0-b7495fcc620a service nova] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Refreshing network info cache for port 07cc827a-876a-4432-a5fc-9ba8920dc5f9 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 980.253542] env[63202]: DEBUG oslo_vmware.api [None req-a456b4ef-c33d-4853-a71d-3eab0fce779d tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385668, 'name': ReconfigVM_Task, 'duration_secs': 0.23233} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.253542] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-a456b4ef-c33d-4853-a71d-3eab0fce779d tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Reconfigured VM instance instance-0000003f to detach disk 2001 {{(pid=63202) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 980.256825] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6c1784fb-f942-4a1f-8b9b-aceee805110e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.278988] env[63202]: DEBUG oslo_vmware.api [None req-a456b4ef-c33d-4853-a71d-3eab0fce779d tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Waiting for the task: (returnval){ [ 980.278988] env[63202]: value = "task-1385670" [ 980.278988] env[63202]: _type = "Task" [ 980.278988] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.300893] env[63202]: DEBUG oslo_vmware.api [None req-a456b4ef-c33d-4853-a71d-3eab0fce779d tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385670, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.310690] env[63202]: DEBUG oslo_vmware.api [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385669, 'name': ReconfigVM_Task, 'duration_secs': 0.363989} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.310941] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Reconfigured VM instance instance-00000056 to attach disk [datastore1] eca51705-a972-48f6-85f5-6c397dad955d/eca51705-a972-48f6-85f5-6c397dad955d.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 980.311608] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5cbb8be5-48b3-425d-9765-82d1342585f7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.313993] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.320798] env[63202]: DEBUG oslo_vmware.api [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 980.320798] env[63202]: value = "task-1385671" [ 980.320798] env[63202]: _type = "Task" [ 980.320798] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.331250] env[63202]: DEBUG oslo_vmware.api [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385671, 'name': Rename_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.488681] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ac13dc1f-36d9-4ead-8a3c-344061e61e29 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquiring lock "7437595c-fa35-483e-95f3-b75405b6bd13" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.488947] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ac13dc1f-36d9-4ead-8a3c-344061e61e29 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "7437595c-fa35-483e-95f3-b75405b6bd13" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.513403] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f2b62a87-5eb1-4d78-93c5-10354cf0ea0e tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.774s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.515612] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.826s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.517240] env[63202]: INFO nova.compute.claims [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 980.533125] env[63202]: INFO nova.scheduler.client.report [None req-f2b62a87-5eb1-4d78-93c5-10354cf0ea0e tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Deleted allocations for instance da285417-bd38-4387-8521-df326ca0e326 [ 980.768969] env[63202]: DEBUG nova.network.neutron [req-7912c24b-66b7-45a9-b387-24c004ab62d3 req-1e8c7953-6023-48bc-acd0-b7495fcc620a service nova] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Updated VIF entry in instance network info cache for port 07cc827a-876a-4432-a5fc-9ba8920dc5f9. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 980.769355] env[63202]: DEBUG nova.network.neutron [req-7912c24b-66b7-45a9-b387-24c004ab62d3 req-1e8c7953-6023-48bc-acd0-b7495fcc620a service nova] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Updating instance_info_cache with network_info: [{"id": "07cc827a-876a-4432-a5fc-9ba8920dc5f9", "address": "fa:16:3e:71:e6:9f", "network": {"id": "06ab5813-9ad9-4021-9bdb-f2f02af8d73f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1714653503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.175", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b2de35030a9484094e964ffc30a822d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3093647a-bab7-4562-ada0-428725e8c0fc", "external-id": "nsx-vlan-transportzone-660", "segmentation_id": 660, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap07cc827a-87", "ovs_interfaceid": "07cc827a-876a-4432-a5fc-9ba8920dc5f9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.776366] env[63202]: DEBUG nova.compute.manager [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 980.790482] env[63202]: DEBUG oslo_vmware.api [None req-a456b4ef-c33d-4853-a71d-3eab0fce779d tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385670, 'name': ReconfigVM_Task, 'duration_secs': 0.170912} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.790930] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-a456b4ef-c33d-4853-a71d-3eab0fce779d tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294155', 'volume_id': '06e1ab70-2c2e-4848-9e08-4e17242e47b5', 'name': 'volume-06e1ab70-2c2e-4848-9e08-4e17242e47b5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb', 'attached_at': '', 'detached_at': '', 'volume_id': '06e1ab70-2c2e-4848-9e08-4e17242e47b5', 'serial': '06e1ab70-2c2e-4848-9e08-4e17242e47b5'} {{(pid=63202) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 980.801471] env[63202]: DEBUG nova.virt.hardware [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 980.801710] env[63202]: DEBUG nova.virt.hardware [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 980.801890] env[63202]: DEBUG nova.virt.hardware [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 980.802100] env[63202]: DEBUG nova.virt.hardware [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 980.802253] env[63202]: DEBUG nova.virt.hardware [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 980.802405] env[63202]: DEBUG nova.virt.hardware [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 980.802620] env[63202]: DEBUG nova.virt.hardware [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 980.802782] env[63202]: DEBUG nova.virt.hardware [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 980.802956] env[63202]: DEBUG nova.virt.hardware [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 980.803152] env[63202]: DEBUG nova.virt.hardware [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 980.803334] env[63202]: DEBUG nova.virt.hardware [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 980.804456] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b17b8c3-e6b1-4780-bd48-c5462eb11b37 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.813252] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b202ea4-f35d-4c0e-978b-8123c8a938f9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.834859] env[63202]: DEBUG oslo_vmware.api [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385671, 'name': Rename_Task, 'duration_secs': 0.185951} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.836026] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 980.836026] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-497c2d00-0fae-4879-ad4d-c4173a50b151 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.842827] env[63202]: DEBUG oslo_vmware.api [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 980.842827] env[63202]: value = "task-1385672" [ 980.842827] env[63202]: _type = "Task" [ 980.842827] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.850641] env[63202]: DEBUG oslo_vmware.api [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385672, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.992592] env[63202]: DEBUG nova.compute.utils [None req-ac13dc1f-36d9-4ead-8a3c-344061e61e29 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 981.042672] env[63202]: DEBUG nova.compute.manager [req-a1f7c93a-c9b5-4489-b830-d9fa0af47ad6 req-8603721b-1c9a-4c31-a09f-50ae40a9c94a service nova] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Received event network-vif-plugged-df7c236b-8be0-46d0-9dc1-56cba5b73ec7 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 981.043178] env[63202]: DEBUG oslo_concurrency.lockutils [req-a1f7c93a-c9b5-4489-b830-d9fa0af47ad6 req-8603721b-1c9a-4c31-a09f-50ae40a9c94a service nova] Acquiring lock "1bfb520e-6708-45a0-bb1f-2037fe1d801e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.043178] env[63202]: DEBUG oslo_concurrency.lockutils [req-a1f7c93a-c9b5-4489-b830-d9fa0af47ad6 req-8603721b-1c9a-4c31-a09f-50ae40a9c94a service nova] Lock "1bfb520e-6708-45a0-bb1f-2037fe1d801e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.043473] env[63202]: DEBUG oslo_concurrency.lockutils [req-a1f7c93a-c9b5-4489-b830-d9fa0af47ad6 req-8603721b-1c9a-4c31-a09f-50ae40a9c94a service nova] Lock "1bfb520e-6708-45a0-bb1f-2037fe1d801e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.043609] env[63202]: DEBUG nova.compute.manager [req-a1f7c93a-c9b5-4489-b830-d9fa0af47ad6 req-8603721b-1c9a-4c31-a09f-50ae40a9c94a service nova] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] No waiting events found dispatching network-vif-plugged-df7c236b-8be0-46d0-9dc1-56cba5b73ec7 {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 981.043808] env[63202]: WARNING nova.compute.manager [req-a1f7c93a-c9b5-4489-b830-d9fa0af47ad6 req-8603721b-1c9a-4c31-a09f-50ae40a9c94a service nova] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Received unexpected event network-vif-plugged-df7c236b-8be0-46d0-9dc1-56cba5b73ec7 for instance with vm_state building and task_state spawning. [ 981.044366] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f2b62a87-5eb1-4d78-93c5-10354cf0ea0e tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lock "da285417-bd38-4387-8521-df326ca0e326" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.835s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.120628] env[63202]: DEBUG nova.network.neutron [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Successfully updated port: df7c236b-8be0-46d0-9dc1-56cba5b73ec7 {{(pid=63202) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 981.271912] env[63202]: DEBUG oslo_concurrency.lockutils [req-7912c24b-66b7-45a9-b387-24c004ab62d3 req-1e8c7953-6023-48bc-acd0-b7495fcc620a service nova] Releasing lock "refresh_cache-8437f856-a707-49c5-b8eb-5a22cdb990f8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.335397] env[63202]: DEBUG nova.objects.instance [None req-a456b4ef-c33d-4853-a71d-3eab0fce779d tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lazy-loading 'flavor' on Instance uuid 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 981.353128] env[63202]: DEBUG oslo_vmware.api [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385672, 'name': PowerOnVM_Task, 'duration_secs': 0.496832} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.354134] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 981.354367] env[63202]: INFO nova.compute.manager [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Took 6.38 seconds to spawn the instance on the hypervisor. [ 981.354554] env[63202]: DEBUG nova.compute.manager [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 981.355408] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61f3fed6-51f4-4137-956b-a1f8dbc9aa53 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.495536] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ac13dc1f-36d9-4ead-8a3c-344061e61e29 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "7437595c-fa35-483e-95f3-b75405b6bd13" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.623155] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "refresh_cache-1bfb520e-6708-45a0-bb1f-2037fe1d801e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.623308] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquired lock "refresh_cache-1bfb520e-6708-45a0-bb1f-2037fe1d801e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.623458] env[63202]: DEBUG nova.network.neutron [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 981.674932] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d615eef0-af43-4981-b15c-6ff1cd1c0ea2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.682679] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89c5255f-2006-4ed4-a864-b5fe2f036816 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.714899] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34b9f0e5-d280-4d15-b0c1-31726f20e408 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.723035] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c77168cd-42bb-482b-9521-09dddafd152a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.736442] env[63202]: DEBUG nova.compute.provider_tree [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 981.875026] env[63202]: INFO nova.compute.manager [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Took 19.17 seconds to build instance. [ 982.056047] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8c9e3966-71f1-4bbb-b3ab-debaeebed328 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquiring lock "7a72a1ab-3b3d-450d-a3a6-572b0035eb0e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.056047] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8c9e3966-71f1-4bbb-b3ab-debaeebed328 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lock "7a72a1ab-3b3d-450d-a3a6-572b0035eb0e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.056047] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8c9e3966-71f1-4bbb-b3ab-debaeebed328 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquiring lock "7a72a1ab-3b3d-450d-a3a6-572b0035eb0e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.056424] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8c9e3966-71f1-4bbb-b3ab-debaeebed328 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lock "7a72a1ab-3b3d-450d-a3a6-572b0035eb0e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.056424] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8c9e3966-71f1-4bbb-b3ab-debaeebed328 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lock "7a72a1ab-3b3d-450d-a3a6-572b0035eb0e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.061271] env[63202]: INFO nova.compute.manager [None req-8c9e3966-71f1-4bbb-b3ab-debaeebed328 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Terminating instance [ 982.063878] env[63202]: DEBUG nova.compute.manager [None req-8c9e3966-71f1-4bbb-b3ab-debaeebed328 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 982.064087] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-8c9e3966-71f1-4bbb-b3ab-debaeebed328 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 982.064941] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55e3d363-5138-4b55-a6e0-d7183afe7440 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.074367] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c9e3966-71f1-4bbb-b3ab-debaeebed328 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 982.075229] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fa44a806-0e2a-4fca-aa21-eefcbd09064c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.084599] env[63202]: DEBUG oslo_vmware.api [None req-8c9e3966-71f1-4bbb-b3ab-debaeebed328 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for the task: (returnval){ [ 982.084599] env[63202]: value = "task-1385673" [ 982.084599] env[63202]: _type = "Task" [ 982.084599] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.093931] env[63202]: DEBUG oslo_vmware.api [None req-8c9e3966-71f1-4bbb-b3ab-debaeebed328 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385673, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.156966] env[63202]: DEBUG nova.network.neutron [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 982.239520] env[63202]: DEBUG nova.scheduler.client.report [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 982.308944] env[63202]: DEBUG nova.network.neutron [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Updating instance_info_cache with network_info: [{"id": "df7c236b-8be0-46d0-9dc1-56cba5b73ec7", "address": "fa:16:3e:93:8e:9b", "network": {"id": "5a3b4c9b-2ca7-4f8b-8bbb-ea000db91402", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-543677750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f508ec04edc844a19640a8a85f27e5b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf7c236b-8b", "ovs_interfaceid": "df7c236b-8be0-46d0-9dc1-56cba5b73ec7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.343244] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a456b4ef-c33d-4853-a71d-3eab0fce779d tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lock "0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.279s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.375974] env[63202]: DEBUG oslo_concurrency.lockutils [None req-43042e16-246f-44cc-afd6-14c94bc46156 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "eca51705-a972-48f6-85f5-6c397dad955d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.685s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.450828] env[63202]: DEBUG oslo_concurrency.lockutils [None req-0be52d89-5eb0-4ffb-85f9-bb2c300ecba1 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Acquiring lock "0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.451152] env[63202]: DEBUG oslo_concurrency.lockutils [None req-0be52d89-5eb0-4ffb-85f9-bb2c300ecba1 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lock "0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.451349] env[63202]: DEBUG nova.compute.manager [None req-0be52d89-5eb0-4ffb-85f9-bb2c300ecba1 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 982.452599] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7bdd988-13c7-4310-96c0-58dc535a95f4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.461054] env[63202]: DEBUG nova.compute.manager [None req-0be52d89-5eb0-4ffb-85f9-bb2c300ecba1 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63202) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 982.461658] env[63202]: DEBUG nova.objects.instance [None req-0be52d89-5eb0-4ffb-85f9-bb2c300ecba1 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lazy-loading 'flavor' on Instance uuid 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 982.522195] env[63202]: DEBUG oslo_concurrency.lockutils [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquiring lock "f23f2c04-c266-46a2-ad3f-34100f6246a0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.522444] env[63202]: DEBUG oslo_concurrency.lockutils [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lock "f23f2c04-c266-46a2-ad3f-34100f6246a0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.564643] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ac13dc1f-36d9-4ead-8a3c-344061e61e29 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquiring lock "7437595c-fa35-483e-95f3-b75405b6bd13" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.564905] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ac13dc1f-36d9-4ead-8a3c-344061e61e29 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "7437595c-fa35-483e-95f3-b75405b6bd13" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.565163] env[63202]: INFO nova.compute.manager [None req-ac13dc1f-36d9-4ead-8a3c-344061e61e29 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Attaching volume d296fff5-0fd4-4820-8a58-c19b5b2d0da9 to /dev/sdb [ 982.594879] env[63202]: DEBUG oslo_vmware.api [None req-8c9e3966-71f1-4bbb-b3ab-debaeebed328 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385673, 'name': PowerOffVM_Task, 'duration_secs': 0.229618} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.595163] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c9e3966-71f1-4bbb-b3ab-debaeebed328 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 982.595334] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-8c9e3966-71f1-4bbb-b3ab-debaeebed328 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 982.595590] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b56ae1b5-93b2-448b-8c95-45bb5f7dbcaf {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.602555] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5441c2c-8177-4cf6-84cc-b0617b8b0d8b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.609829] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8705e12a-974e-4663-b92b-548f92955ab5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.646895] env[63202]: DEBUG nova.virt.block_device [None req-ac13dc1f-36d9-4ead-8a3c-344061e61e29 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Updating existing volume attachment record: beeb43f5-1cbb-47bf-b733-d74413bb740e {{(pid=63202) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 982.701260] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-8c9e3966-71f1-4bbb-b3ab-debaeebed328 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 982.701436] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-8c9e3966-71f1-4bbb-b3ab-debaeebed328 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Deleting contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 982.701744] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-8c9e3966-71f1-4bbb-b3ab-debaeebed328 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Deleting the datastore file [datastore1] 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 982.701984] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ff69b89c-dd26-406e-878c-c4a544f3d709 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.710446] env[63202]: DEBUG oslo_vmware.api [None req-8c9e3966-71f1-4bbb-b3ab-debaeebed328 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for the task: (returnval){ [ 982.710446] env[63202]: value = "task-1385675" [ 982.710446] env[63202]: _type = "Task" [ 982.710446] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.719276] env[63202]: DEBUG oslo_vmware.api [None req-8c9e3966-71f1-4bbb-b3ab-debaeebed328 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385675, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.744371] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.229s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.744879] env[63202]: DEBUG nova.compute.manager [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 982.747638] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e6c16b26-535e-43d0-9def-02a7399b5a99 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.577s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.747728] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e6c16b26-535e-43d0-9def-02a7399b5a99 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.749687] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.436s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.751471] env[63202]: INFO nova.compute.claims [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 982.775653] env[63202]: INFO nova.scheduler.client.report [None req-e6c16b26-535e-43d0-9def-02a7399b5a99 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Deleted allocations for instance e775e5e1-521a-4fc7-80e6-bcb6a70516c5 [ 982.811699] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Releasing lock "refresh_cache-1bfb520e-6708-45a0-bb1f-2037fe1d801e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.812065] env[63202]: DEBUG nova.compute.manager [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Instance network_info: |[{"id": "df7c236b-8be0-46d0-9dc1-56cba5b73ec7", "address": "fa:16:3e:93:8e:9b", "network": {"id": "5a3b4c9b-2ca7-4f8b-8bbb-ea000db91402", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-543677750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f508ec04edc844a19640a8a85f27e5b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf7c236b-8b", "ovs_interfaceid": "df7c236b-8be0-46d0-9dc1-56cba5b73ec7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 982.812801] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:93:8e:9b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'df7c236b-8be0-46d0-9dc1-56cba5b73ec7', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 982.820417] env[63202]: DEBUG oslo.service.loopingcall [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 982.821038] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 982.821240] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-858d6b12-2012-4207-a71c-5e301a563dd1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.842201] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 982.842201] env[63202]: value = "task-1385677" [ 982.842201] env[63202]: _type = "Task" [ 982.842201] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.850455] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385677, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.968214] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-0be52d89-5eb0-4ffb-85f9-bb2c300ecba1 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 982.968812] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2e6f0de8-2be5-4fae-9129-03659758b4ab {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.976230] env[63202]: DEBUG oslo_vmware.api [None req-0be52d89-5eb0-4ffb-85f9-bb2c300ecba1 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Waiting for the task: (returnval){ [ 982.976230] env[63202]: value = "task-1385680" [ 982.976230] env[63202]: _type = "Task" [ 982.976230] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.985173] env[63202]: DEBUG oslo_vmware.api [None req-0be52d89-5eb0-4ffb-85f9-bb2c300ecba1 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385680, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.024999] env[63202]: DEBUG nova.compute.manager [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 983.069166] env[63202]: DEBUG nova.compute.manager [req-e510a0c1-5c42-499b-8167-08c74c3015b0 req-55e995ee-83ee-4f45-890a-2a580f6bde74 service nova] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Received event network-changed-df7c236b-8be0-46d0-9dc1-56cba5b73ec7 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 983.069978] env[63202]: DEBUG nova.compute.manager [req-e510a0c1-5c42-499b-8167-08c74c3015b0 req-55e995ee-83ee-4f45-890a-2a580f6bde74 service nova] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Refreshing instance network info cache due to event network-changed-df7c236b-8be0-46d0-9dc1-56cba5b73ec7. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 983.069978] env[63202]: DEBUG oslo_concurrency.lockutils [req-e510a0c1-5c42-499b-8167-08c74c3015b0 req-55e995ee-83ee-4f45-890a-2a580f6bde74 service nova] Acquiring lock "refresh_cache-1bfb520e-6708-45a0-bb1f-2037fe1d801e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.069978] env[63202]: DEBUG oslo_concurrency.lockutils [req-e510a0c1-5c42-499b-8167-08c74c3015b0 req-55e995ee-83ee-4f45-890a-2a580f6bde74 service nova] Acquired lock "refresh_cache-1bfb520e-6708-45a0-bb1f-2037fe1d801e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.069978] env[63202]: DEBUG nova.network.neutron [req-e510a0c1-5c42-499b-8167-08c74c3015b0 req-55e995ee-83ee-4f45-890a-2a580f6bde74 service nova] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Refreshing network info cache for port df7c236b-8be0-46d0-9dc1-56cba5b73ec7 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 983.133775] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8b29a4d0-9d52-4d7e-8f06-7f965b6903bb tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "eca51705-a972-48f6-85f5-6c397dad955d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.134056] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8b29a4d0-9d52-4d7e-8f06-7f965b6903bb tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "eca51705-a972-48f6-85f5-6c397dad955d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.134350] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8b29a4d0-9d52-4d7e-8f06-7f965b6903bb tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "eca51705-a972-48f6-85f5-6c397dad955d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.134585] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8b29a4d0-9d52-4d7e-8f06-7f965b6903bb tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "eca51705-a972-48f6-85f5-6c397dad955d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.135218] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8b29a4d0-9d52-4d7e-8f06-7f965b6903bb tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "eca51705-a972-48f6-85f5-6c397dad955d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.137489] env[63202]: INFO nova.compute.manager [None req-8b29a4d0-9d52-4d7e-8f06-7f965b6903bb tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Terminating instance [ 983.142018] env[63202]: DEBUG nova.compute.manager [None req-8b29a4d0-9d52-4d7e-8f06-7f965b6903bb tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 983.142018] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-8b29a4d0-9d52-4d7e-8f06-7f965b6903bb tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 983.142018] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e44c74c8-cebe-4e81-9653-229efb73b763 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.152948] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b29a4d0-9d52-4d7e-8f06-7f965b6903bb tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 983.154290] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ec67b2fd-8fd8-45e9-95ec-c38f8eb41f96 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.161214] env[63202]: DEBUG oslo_vmware.api [None req-8b29a4d0-9d52-4d7e-8f06-7f965b6903bb tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 983.161214] env[63202]: value = "task-1385681" [ 983.161214] env[63202]: _type = "Task" [ 983.161214] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.172132] env[63202]: DEBUG oslo_vmware.api [None req-8b29a4d0-9d52-4d7e-8f06-7f965b6903bb tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385681, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.222816] env[63202]: DEBUG oslo_vmware.api [None req-8c9e3966-71f1-4bbb-b3ab-debaeebed328 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385675, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.256582] env[63202]: DEBUG nova.compute.utils [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 983.262393] env[63202]: DEBUG nova.compute.manager [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 983.262742] env[63202]: DEBUG nova.network.neutron [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 983.285146] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e6c16b26-535e-43d0-9def-02a7399b5a99 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "e775e5e1-521a-4fc7-80e6-bcb6a70516c5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.203s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.323931] env[63202]: DEBUG nova.policy [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1236175893a3401cb01b04604168f1e0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '22352c25bca8416a948014391a5389ac', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 983.355577] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385677, 'name': CreateVM_Task, 'duration_secs': 0.444268} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.356741] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 983.358715] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.358715] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.358715] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 983.358715] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5a97785d-8856-48a9-934b-857d57f41998 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.365141] env[63202]: DEBUG oslo_vmware.api [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 983.365141] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]522c7761-ec01-e47f-d5cd-35758a44d80b" [ 983.365141] env[63202]: _type = "Task" [ 983.365141] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.378114] env[63202]: DEBUG oslo_vmware.api [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]522c7761-ec01-e47f-d5cd-35758a44d80b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.415512] env[63202]: DEBUG oslo_concurrency.lockutils [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Acquiring lock "0d7fd297-37db-4876-bb91-5bb0b7dc335e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.416671] env[63202]: DEBUG oslo_concurrency.lockutils [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Lock "0d7fd297-37db-4876-bb91-5bb0b7dc335e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.486729] env[63202]: DEBUG oslo_vmware.api [None req-0be52d89-5eb0-4ffb-85f9-bb2c300ecba1 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385680, 'name': PowerOffVM_Task, 'duration_secs': 0.209378} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.487934] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-0be52d89-5eb0-4ffb-85f9-bb2c300ecba1 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 983.487934] env[63202]: DEBUG nova.compute.manager [None req-0be52d89-5eb0-4ffb-85f9-bb2c300ecba1 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 983.488146] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a7f6473-f5bc-4a33-b4c4-49de01663607 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.548856] env[63202]: DEBUG oslo_concurrency.lockutils [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.601254] env[63202]: DEBUG nova.network.neutron [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Successfully created port: 0bd2810b-f01e-4cd8-9c3e-76ac2339047a {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 983.672535] env[63202]: DEBUG oslo_vmware.api [None req-8b29a4d0-9d52-4d7e-8f06-7f965b6903bb tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385681, 'name': PowerOffVM_Task, 'duration_secs': 0.217883} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.672762] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b29a4d0-9d52-4d7e-8f06-7f965b6903bb tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 983.672936] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-8b29a4d0-9d52-4d7e-8f06-7f965b6903bb tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 983.673206] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b90cb421-7dbd-4a56-b19c-dc91b6493b6e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.724114] env[63202]: DEBUG oslo_vmware.api [None req-8c9e3966-71f1-4bbb-b3ab-debaeebed328 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385675, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.722562} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.724413] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-8c9e3966-71f1-4bbb-b3ab-debaeebed328 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 983.724606] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-8c9e3966-71f1-4bbb-b3ab-debaeebed328 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Deleted contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 983.724786] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-8c9e3966-71f1-4bbb-b3ab-debaeebed328 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 983.724955] env[63202]: INFO nova.compute.manager [None req-8c9e3966-71f1-4bbb-b3ab-debaeebed328 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Took 1.66 seconds to destroy the instance on the hypervisor. [ 983.725215] env[63202]: DEBUG oslo.service.loopingcall [None req-8c9e3966-71f1-4bbb-b3ab-debaeebed328 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 983.725412] env[63202]: DEBUG nova.compute.manager [-] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 983.725504] env[63202]: DEBUG nova.network.neutron [-] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 983.742037] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-8b29a4d0-9d52-4d7e-8f06-7f965b6903bb tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 983.742292] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-8b29a4d0-9d52-4d7e-8f06-7f965b6903bb tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Deleting contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 983.742483] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b29a4d0-9d52-4d7e-8f06-7f965b6903bb tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Deleting the datastore file [datastore1] eca51705-a972-48f6-85f5-6c397dad955d {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 983.742838] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3da1e315-62e7-46d5-b21e-5e36cdbbb1ac {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.750361] env[63202]: DEBUG oslo_vmware.api [None req-8b29a4d0-9d52-4d7e-8f06-7f965b6903bb tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 983.750361] env[63202]: value = "task-1385683" [ 983.750361] env[63202]: _type = "Task" [ 983.750361] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.761035] env[63202]: DEBUG oslo_vmware.api [None req-8b29a4d0-9d52-4d7e-8f06-7f965b6903bb tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385683, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.765287] env[63202]: DEBUG nova.compute.manager [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 983.812241] env[63202]: DEBUG nova.network.neutron [req-e510a0c1-5c42-499b-8167-08c74c3015b0 req-55e995ee-83ee-4f45-890a-2a580f6bde74 service nova] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Updated VIF entry in instance network info cache for port df7c236b-8be0-46d0-9dc1-56cba5b73ec7. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 983.812652] env[63202]: DEBUG nova.network.neutron [req-e510a0c1-5c42-499b-8167-08c74c3015b0 req-55e995ee-83ee-4f45-890a-2a580f6bde74 service nova] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Updating instance_info_cache with network_info: [{"id": "df7c236b-8be0-46d0-9dc1-56cba5b73ec7", "address": "fa:16:3e:93:8e:9b", "network": {"id": "5a3b4c9b-2ca7-4f8b-8bbb-ea000db91402", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-543677750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f508ec04edc844a19640a8a85f27e5b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf7c236b-8b", "ovs_interfaceid": "df7c236b-8be0-46d0-9dc1-56cba5b73ec7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.878858] env[63202]: DEBUG oslo_vmware.api [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]522c7761-ec01-e47f-d5cd-35758a44d80b, 'name': SearchDatastore_Task, 'duration_secs': 0.024014} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.881873] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.882295] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 983.882653] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.882927] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.883236] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 983.884515] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-be450466-48f8-45f6-9ed1-00a89be6d33e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.900757] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 983.900963] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 983.901778] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4291dc5b-19ed-4a68-b4a6-a252dcc6a6b5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.911035] env[63202]: DEBUG oslo_vmware.api [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 983.911035] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52478402-fc45-5499-b8b3-3a9b1ef6491a" [ 983.911035] env[63202]: _type = "Task" [ 983.911035] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.920749] env[63202]: DEBUG nova.compute.manager [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 983.928849] env[63202]: DEBUG oslo_vmware.api [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52478402-fc45-5499-b8b3-3a9b1ef6491a, 'name': SearchDatastore_Task, 'duration_secs': 0.01101} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.932176] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-284eee45-4f21-4036-af4c-0ff21021e058 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.938376] env[63202]: DEBUG oslo_vmware.api [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 983.938376] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]523247f1-b311-e321-9556-2675756a9d91" [ 983.938376] env[63202]: _type = "Task" [ 983.938376] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.947786] env[63202]: DEBUG oslo_vmware.api [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]523247f1-b311-e321-9556-2675756a9d91, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.958087] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5878c25a-e1a2-46ac-8ae2-20f71bc64ad9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.966270] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbfad031-7ed4-4990-ae6c-5806c4033e45 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.998427] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d4e01bf-9733-4baf-bf20-1c4eb3179919 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.005655] env[63202]: DEBUG oslo_concurrency.lockutils [None req-0be52d89-5eb0-4ffb-85f9-bb2c300ecba1 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lock "0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.554s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.010518] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7b126bc-b3e5-4da6-9917-11c6237887fd {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.024954] env[63202]: DEBUG nova.compute.provider_tree [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 984.261488] env[63202]: DEBUG oslo_vmware.api [None req-8b29a4d0-9d52-4d7e-8f06-7f965b6903bb tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385683, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.315314] env[63202]: DEBUG oslo_concurrency.lockutils [req-e510a0c1-5c42-499b-8167-08c74c3015b0 req-55e995ee-83ee-4f45-890a-2a580f6bde74 service nova] Releasing lock "refresh_cache-1bfb520e-6708-45a0-bb1f-2037fe1d801e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.444282] env[63202]: DEBUG oslo_concurrency.lockutils [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.452304] env[63202]: DEBUG oslo_vmware.api [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]523247f1-b311-e321-9556-2675756a9d91, 'name': SearchDatastore_Task, 'duration_secs': 0.017701} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.452552] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.452815] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] 1bfb520e-6708-45a0-bb1f-2037fe1d801e/1bfb520e-6708-45a0-bb1f-2037fe1d801e.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 984.453098] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-04ef52da-9c70-46fc-99f9-c74c89211aa0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.461255] env[63202]: DEBUG oslo_vmware.api [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 984.461255] env[63202]: value = "task-1385684" [ 984.461255] env[63202]: _type = "Task" [ 984.461255] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.469230] env[63202]: DEBUG oslo_vmware.api [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385684, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.527825] env[63202]: DEBUG nova.scheduler.client.report [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 984.560223] env[63202]: DEBUG nova.network.neutron [-] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 984.763879] env[63202]: DEBUG oslo_vmware.api [None req-8b29a4d0-9d52-4d7e-8f06-7f965b6903bb tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385683, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.525474} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.764170] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b29a4d0-9d52-4d7e-8f06-7f965b6903bb tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 984.764360] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-8b29a4d0-9d52-4d7e-8f06-7f965b6903bb tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Deleted contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 984.764542] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-8b29a4d0-9d52-4d7e-8f06-7f965b6903bb tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 984.764721] env[63202]: INFO nova.compute.manager [None req-8b29a4d0-9d52-4d7e-8f06-7f965b6903bb tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Took 1.62 seconds to destroy the instance on the hypervisor. [ 984.764968] env[63202]: DEBUG oslo.service.loopingcall [None req-8b29a4d0-9d52-4d7e-8f06-7f965b6903bb tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 984.765190] env[63202]: DEBUG nova.compute.manager [-] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 984.765284] env[63202]: DEBUG nova.network.neutron [-] [instance: eca51705-a972-48f6-85f5-6c397dad955d] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 984.780026] env[63202]: DEBUG nova.compute.manager [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 984.807376] env[63202]: DEBUG nova.virt.hardware [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 984.807679] env[63202]: DEBUG nova.virt.hardware [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 984.807900] env[63202]: DEBUG nova.virt.hardware [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 984.808139] env[63202]: DEBUG nova.virt.hardware [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 984.808284] env[63202]: DEBUG nova.virt.hardware [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 984.808490] env[63202]: DEBUG nova.virt.hardware [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 984.808676] env[63202]: DEBUG nova.virt.hardware [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 984.808922] env[63202]: DEBUG nova.virt.hardware [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 984.809177] env[63202]: DEBUG nova.virt.hardware [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 984.809478] env[63202]: DEBUG nova.virt.hardware [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 984.809637] env[63202]: DEBUG nova.virt.hardware [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 984.810620] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b13aad4-d17a-4705-a57b-39d89ca0f0fb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.819511] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25159a2d-6337-48cf-b78a-c3fb275d9406 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.975755] env[63202]: DEBUG oslo_vmware.api [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385684, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.033907] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.283s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.033907] env[63202]: DEBUG nova.compute.manager [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 985.040019] env[63202]: DEBUG oslo_concurrency.lockutils [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.488s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.040019] env[63202]: INFO nova.compute.claims [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 985.064475] env[63202]: INFO nova.compute.manager [-] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Took 1.34 seconds to deallocate network for instance. [ 985.098851] env[63202]: DEBUG nova.compute.manager [req-e2bf9c55-cd8e-4a4d-9cfa-92864cfec9c3 req-7d309637-ce5f-4799-997e-8f259b71b74e service nova] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Received event network-vif-deleted-aa2cbb5d-3724-4754-922b-da0798168b84 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 985.194106] env[63202]: DEBUG nova.objects.instance [None req-95d4c183-16f0-4fd7-bdde-4f0d2160b43f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lazy-loading 'flavor' on Instance uuid 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 985.298429] env[63202]: DEBUG nova.network.neutron [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Successfully updated port: 0bd2810b-f01e-4cd8-9c3e-76ac2339047a {{(pid=63202) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 985.471983] env[63202]: DEBUG oslo_vmware.api [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385684, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.560973} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.473048] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] 1bfb520e-6708-45a0-bb1f-2037fe1d801e/1bfb520e-6708-45a0-bb1f-2037fe1d801e.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 985.473048] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 985.473048] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-20d9783d-3c6e-433e-9fe2-ee55652050a2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.479884] env[63202]: DEBUG oslo_vmware.api [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 985.479884] env[63202]: value = "task-1385686" [ 985.479884] env[63202]: _type = "Task" [ 985.479884] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.487767] env[63202]: DEBUG oslo_vmware.api [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385686, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.542416] env[63202]: DEBUG nova.compute.utils [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 985.545600] env[63202]: DEBUG nova.compute.manager [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 985.545796] env[63202]: DEBUG nova.network.neutron [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 985.571838] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8c9e3966-71f1-4bbb-b3ab-debaeebed328 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.586574] env[63202]: DEBUG nova.network.neutron [-] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.591059] env[63202]: DEBUG nova.policy [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2b9d1001300c476a896f335b3acf041b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f651d18a9559499da27f301ee75bb2fb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 985.699149] env[63202]: DEBUG oslo_concurrency.lockutils [None req-95d4c183-16f0-4fd7-bdde-4f0d2160b43f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Acquiring lock "refresh_cache-0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.699382] env[63202]: DEBUG oslo_concurrency.lockutils [None req-95d4c183-16f0-4fd7-bdde-4f0d2160b43f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Acquired lock "refresh_cache-0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.699496] env[63202]: DEBUG nova.network.neutron [None req-95d4c183-16f0-4fd7-bdde-4f0d2160b43f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 985.699674] env[63202]: DEBUG nova.objects.instance [None req-95d4c183-16f0-4fd7-bdde-4f0d2160b43f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lazy-loading 'info_cache' on Instance uuid 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 985.806145] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "refresh_cache-dd146c71-c391-41e8-8cc7-7276f4a518d9" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.806145] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquired lock "refresh_cache-dd146c71-c391-41e8-8cc7-7276f4a518d9" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.806145] env[63202]: DEBUG nova.network.neutron [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 985.833378] env[63202]: DEBUG nova.network.neutron [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Successfully created port: e7f10c86-d09d-4e51-a8f7-deef174f2539 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 985.990802] env[63202]: DEBUG oslo_vmware.api [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385686, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062259} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.990802] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 985.991082] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86ce743b-d42e-4776-b135-76dd620078d3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.013158] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Reconfiguring VM instance instance-00000057 to attach disk [datastore1] 1bfb520e-6708-45a0-bb1f-2037fe1d801e/1bfb520e-6708-45a0-bb1f-2037fe1d801e.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 986.013429] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-64cbd697-b6d4-4443-bfd6-dea0f341e08f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.034212] env[63202]: DEBUG oslo_vmware.api [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 986.034212] env[63202]: value = "task-1385687" [ 986.034212] env[63202]: _type = "Task" [ 986.034212] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.042848] env[63202]: DEBUG oslo_vmware.api [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385687, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.046490] env[63202]: DEBUG nova.compute.manager [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 986.088714] env[63202]: INFO nova.compute.manager [-] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Took 1.32 seconds to deallocate network for instance. [ 986.202892] env[63202]: DEBUG nova.objects.base [None req-95d4c183-16f0-4fd7-bdde-4f0d2160b43f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Object Instance<0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb> lazy-loaded attributes: flavor,info_cache {{(pid=63202) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 986.210712] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fab51323-1ddf-4269-a9a8-34bd9255a45c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.220405] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11758bd6-3415-4a11-b44b-9055867a3809 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.257894] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90cf59eb-806f-4343-9a7a-99fec644a370 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.264867] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa3da485-4467-43d8-82b9-c35ee8f6481f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.281099] env[63202]: DEBUG nova.compute.provider_tree [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 986.348223] env[63202]: DEBUG nova.network.neutron [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 986.508992] env[63202]: DEBUG nova.network.neutron [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Updating instance_info_cache with network_info: [{"id": "0bd2810b-f01e-4cd8-9c3e-76ac2339047a", "address": "fa:16:3e:9f:08:72", "network": {"id": "18390479-16ce-4012-9ba6-abf19b5616d1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1376026898-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22352c25bca8416a948014391a5389ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "085fb0ff-9285-4f1d-a008-a14da4844357", "external-id": "nsx-vlan-transportzone-729", "segmentation_id": 729, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0bd2810b-f0", "ovs_interfaceid": "0bd2810b-f01e-4cd8-9c3e-76ac2339047a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.544015] env[63202]: DEBUG oslo_vmware.api [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385687, 'name': ReconfigVM_Task, 'duration_secs': 0.319029} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.544361] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Reconfigured VM instance instance-00000057 to attach disk [datastore1] 1bfb520e-6708-45a0-bb1f-2037fe1d801e/1bfb520e-6708-45a0-bb1f-2037fe1d801e.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 986.545055] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e07555d7-5ab3-43b5-9b27-dc3077e7bc4d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.556589] env[63202]: DEBUG oslo_vmware.api [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 986.556589] env[63202]: value = "task-1385688" [ 986.556589] env[63202]: _type = "Task" [ 986.556589] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.566558] env[63202]: DEBUG oslo_vmware.api [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385688, 'name': Rename_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.595861] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8b29a4d0-9d52-4d7e-8f06-7f965b6903bb tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.784083] env[63202]: DEBUG nova.scheduler.client.report [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 986.905313] env[63202]: DEBUG nova.network.neutron [None req-95d4c183-16f0-4fd7-bdde-4f0d2160b43f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Updating instance_info_cache with network_info: [{"id": "e5621970-d0a8-4803-94ef-f6d577bdf6d2", "address": "fa:16:3e:00:ba:72", "network": {"id": "345538f4-e791-4fc6-9719-237c11f1382a", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-449651008-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.170", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a9e939cdf4a94a8aa0107a5761771c2c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa09e855-8af1-419b-b78d-8ffcc94b1bfb", "external-id": "nsx-vlan-transportzone-901", "segmentation_id": 901, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape5621970-d0", "ovs_interfaceid": "e5621970-d0a8-4803-94ef-f6d577bdf6d2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.012913] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Releasing lock "refresh_cache-dd146c71-c391-41e8-8cc7-7276f4a518d9" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.013336] env[63202]: DEBUG nova.compute.manager [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Instance network_info: |[{"id": "0bd2810b-f01e-4cd8-9c3e-76ac2339047a", "address": "fa:16:3e:9f:08:72", "network": {"id": "18390479-16ce-4012-9ba6-abf19b5616d1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1376026898-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22352c25bca8416a948014391a5389ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "085fb0ff-9285-4f1d-a008-a14da4844357", "external-id": "nsx-vlan-transportzone-729", "segmentation_id": 729, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0bd2810b-f0", "ovs_interfaceid": "0bd2810b-f01e-4cd8-9c3e-76ac2339047a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 987.013805] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9f:08:72', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '085fb0ff-9285-4f1d-a008-a14da4844357', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0bd2810b-f01e-4cd8-9c3e-76ac2339047a', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 987.021497] env[63202]: DEBUG oslo.service.loopingcall [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 987.022159] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 987.022389] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-526c0c86-3af9-4147-979f-f43ec0e07c53 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.044253] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 987.044253] env[63202]: value = "task-1385689" [ 987.044253] env[63202]: _type = "Task" [ 987.044253] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.052554] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385689, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.061927] env[63202]: DEBUG nova.compute.manager [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 987.067142] env[63202]: DEBUG oslo_vmware.api [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385688, 'name': Rename_Task, 'duration_secs': 0.141007} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.068130] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 987.068130] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-af2c2e16-cb31-415d-8573-35abd842b090 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.075347] env[63202]: DEBUG oslo_vmware.api [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 987.075347] env[63202]: value = "task-1385690" [ 987.075347] env[63202]: _type = "Task" [ 987.075347] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.086391] env[63202]: DEBUG oslo_vmware.api [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385690, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.096483] env[63202]: DEBUG nova.virt.hardware [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 987.096793] env[63202]: DEBUG nova.virt.hardware [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 987.096984] env[63202]: DEBUG nova.virt.hardware [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 987.097285] env[63202]: DEBUG nova.virt.hardware [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 987.097526] env[63202]: DEBUG nova.virt.hardware [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 987.097703] env[63202]: DEBUG nova.virt.hardware [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 987.097986] env[63202]: DEBUG nova.virt.hardware [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 987.098079] env[63202]: DEBUG nova.virt.hardware [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 987.098256] env[63202]: DEBUG nova.virt.hardware [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 987.098436] env[63202]: DEBUG nova.virt.hardware [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 987.098652] env[63202]: DEBUG nova.virt.hardware [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 987.099518] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95dbd13d-68c0-4e71-937c-5386f6570419 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.107972] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf580d28-8cdf-4bb4-9d5e-7db8f69faf34 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.129134] env[63202]: DEBUG nova.compute.manager [req-eea299f2-babc-4ed7-9b3e-5739530a4e42 req-be703a76-fe50-4821-b24e-2235e681840c service nova] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Received event network-vif-deleted-145b6d77-5716-400a-911a-37decf12fae3 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 987.129349] env[63202]: DEBUG nova.compute.manager [req-eea299f2-babc-4ed7-9b3e-5739530a4e42 req-be703a76-fe50-4821-b24e-2235e681840c service nova] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Received event network-vif-plugged-0bd2810b-f01e-4cd8-9c3e-76ac2339047a {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 987.129523] env[63202]: DEBUG oslo_concurrency.lockutils [req-eea299f2-babc-4ed7-9b3e-5739530a4e42 req-be703a76-fe50-4821-b24e-2235e681840c service nova] Acquiring lock "dd146c71-c391-41e8-8cc7-7276f4a518d9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.129706] env[63202]: DEBUG oslo_concurrency.lockutils [req-eea299f2-babc-4ed7-9b3e-5739530a4e42 req-be703a76-fe50-4821-b24e-2235e681840c service nova] Lock "dd146c71-c391-41e8-8cc7-7276f4a518d9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.129873] env[63202]: DEBUG oslo_concurrency.lockutils [req-eea299f2-babc-4ed7-9b3e-5739530a4e42 req-be703a76-fe50-4821-b24e-2235e681840c service nova] Lock "dd146c71-c391-41e8-8cc7-7276f4a518d9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.130048] env[63202]: DEBUG nova.compute.manager [req-eea299f2-babc-4ed7-9b3e-5739530a4e42 req-be703a76-fe50-4821-b24e-2235e681840c service nova] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] No waiting events found dispatching network-vif-plugged-0bd2810b-f01e-4cd8-9c3e-76ac2339047a {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 987.130217] env[63202]: WARNING nova.compute.manager [req-eea299f2-babc-4ed7-9b3e-5739530a4e42 req-be703a76-fe50-4821-b24e-2235e681840c service nova] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Received unexpected event network-vif-plugged-0bd2810b-f01e-4cd8-9c3e-76ac2339047a for instance with vm_state building and task_state spawning. [ 987.130377] env[63202]: DEBUG nova.compute.manager [req-eea299f2-babc-4ed7-9b3e-5739530a4e42 req-be703a76-fe50-4821-b24e-2235e681840c service nova] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Received event network-changed-0bd2810b-f01e-4cd8-9c3e-76ac2339047a {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 987.130604] env[63202]: DEBUG nova.compute.manager [req-eea299f2-babc-4ed7-9b3e-5739530a4e42 req-be703a76-fe50-4821-b24e-2235e681840c service nova] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Refreshing instance network info cache due to event network-changed-0bd2810b-f01e-4cd8-9c3e-76ac2339047a. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 987.130738] env[63202]: DEBUG oslo_concurrency.lockutils [req-eea299f2-babc-4ed7-9b3e-5739530a4e42 req-be703a76-fe50-4821-b24e-2235e681840c service nova] Acquiring lock "refresh_cache-dd146c71-c391-41e8-8cc7-7276f4a518d9" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.130878] env[63202]: DEBUG oslo_concurrency.lockutils [req-eea299f2-babc-4ed7-9b3e-5739530a4e42 req-be703a76-fe50-4821-b24e-2235e681840c service nova] Acquired lock "refresh_cache-dd146c71-c391-41e8-8cc7-7276f4a518d9" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.131041] env[63202]: DEBUG nova.network.neutron [req-eea299f2-babc-4ed7-9b3e-5739530a4e42 req-be703a76-fe50-4821-b24e-2235e681840c service nova] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Refreshing network info cache for port 0bd2810b-f01e-4cd8-9c3e-76ac2339047a {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 987.197301] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-ac13dc1f-36d9-4ead-8a3c-344061e61e29 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Volume attach. Driver type: vmdk {{(pid=63202) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 987.197534] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-ac13dc1f-36d9-4ead-8a3c-344061e61e29 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294189', 'volume_id': 'd296fff5-0fd4-4820-8a58-c19b5b2d0da9', 'name': 'volume-d296fff5-0fd4-4820-8a58-c19b5b2d0da9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7437595c-fa35-483e-95f3-b75405b6bd13', 'attached_at': '', 'detached_at': '', 'volume_id': 'd296fff5-0fd4-4820-8a58-c19b5b2d0da9', 'serial': 'd296fff5-0fd4-4820-8a58-c19b5b2d0da9'} {{(pid=63202) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 987.198795] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65304c8a-8cc1-4427-83e4-554e82c766cd {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.216698] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cb85237-c477-420d-aaf7-230d34cd1699 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.245915] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-ac13dc1f-36d9-4ead-8a3c-344061e61e29 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Reconfiguring VM instance instance-00000051 to attach disk [datastore1] volume-d296fff5-0fd4-4820-8a58-c19b5b2d0da9/volume-d296fff5-0fd4-4820-8a58-c19b5b2d0da9.vmdk or device None with type thin {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 987.246460] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cb976432-4672-4834-9e3d-835910e4fe0e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.266091] env[63202]: DEBUG oslo_concurrency.lockutils [None req-eefe30e7-4320-41e9-881f-1b00135a970e tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Acquiring lock "f3128c09-3680-4b0b-b463-3d6cd203fcf4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.266091] env[63202]: DEBUG oslo_concurrency.lockutils [None req-eefe30e7-4320-41e9-881f-1b00135a970e tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Lock "f3128c09-3680-4b0b-b463-3d6cd203fcf4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.266091] env[63202]: DEBUG oslo_concurrency.lockutils [None req-eefe30e7-4320-41e9-881f-1b00135a970e tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Acquiring lock "f3128c09-3680-4b0b-b463-3d6cd203fcf4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.266091] env[63202]: DEBUG oslo_concurrency.lockutils [None req-eefe30e7-4320-41e9-881f-1b00135a970e tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Lock "f3128c09-3680-4b0b-b463-3d6cd203fcf4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.266306] env[63202]: DEBUG oslo_concurrency.lockutils [None req-eefe30e7-4320-41e9-881f-1b00135a970e tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Lock "f3128c09-3680-4b0b-b463-3d6cd203fcf4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.266545] env[63202]: INFO nova.compute.manager [None req-eefe30e7-4320-41e9-881f-1b00135a970e tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Terminating instance [ 987.268995] env[63202]: DEBUG nova.compute.manager [None req-eefe30e7-4320-41e9-881f-1b00135a970e tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 987.269285] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-eefe30e7-4320-41e9-881f-1b00135a970e tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 987.270091] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5e14fa0-0f16-4863-8e40-45f0843d97c3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.274997] env[63202]: DEBUG oslo_vmware.api [None req-ac13dc1f-36d9-4ead-8a3c-344061e61e29 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 987.274997] env[63202]: value = "task-1385691" [ 987.274997] env[63202]: _type = "Task" [ 987.274997] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.280374] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-eefe30e7-4320-41e9-881f-1b00135a970e tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 987.281027] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8f2596f9-7bab-4f9f-a6c2-5ecd55fc4c24 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.286191] env[63202]: DEBUG oslo_vmware.api [None req-ac13dc1f-36d9-4ead-8a3c-344061e61e29 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385691, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.287446] env[63202]: DEBUG oslo_vmware.api [None req-eefe30e7-4320-41e9-881f-1b00135a970e tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Waiting for the task: (returnval){ [ 987.287446] env[63202]: value = "task-1385692" [ 987.287446] env[63202]: _type = "Task" [ 987.287446] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.290852] env[63202]: DEBUG oslo_concurrency.lockutils [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.255s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.291365] env[63202]: DEBUG nova.compute.manager [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 987.294293] env[63202]: DEBUG oslo_concurrency.lockutils [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.850s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.295811] env[63202]: INFO nova.compute.claims [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 987.304864] env[63202]: DEBUG oslo_vmware.api [None req-eefe30e7-4320-41e9-881f-1b00135a970e tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Task: {'id': task-1385692, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.339774] env[63202]: DEBUG nova.network.neutron [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Successfully updated port: e7f10c86-d09d-4e51-a8f7-deef174f2539 {{(pid=63202) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 987.410039] env[63202]: DEBUG oslo_concurrency.lockutils [None req-95d4c183-16f0-4fd7-bdde-4f0d2160b43f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Releasing lock "refresh_cache-0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.558993] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385689, 'name': CreateVM_Task} progress is 25%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.586973] env[63202]: DEBUG oslo_vmware.api [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385690, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.785160] env[63202]: DEBUG oslo_vmware.api [None req-ac13dc1f-36d9-4ead-8a3c-344061e61e29 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385691, 'name': ReconfigVM_Task, 'duration_secs': 0.390621} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.785450] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-ac13dc1f-36d9-4ead-8a3c-344061e61e29 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Reconfigured VM instance instance-00000051 to attach disk [datastore1] volume-d296fff5-0fd4-4820-8a58-c19b5b2d0da9/volume-d296fff5-0fd4-4820-8a58-c19b5b2d0da9.vmdk or device None with type thin {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 987.790199] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-04cc6082-743c-46d4-b013-4f0f7327b61e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.803350] env[63202]: DEBUG nova.compute.utils [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 987.806650] env[63202]: DEBUG nova.compute.manager [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 987.806812] env[63202]: DEBUG nova.network.neutron [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 987.817573] env[63202]: DEBUG oslo_vmware.api [None req-eefe30e7-4320-41e9-881f-1b00135a970e tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Task: {'id': task-1385692, 'name': PowerOffVM_Task, 'duration_secs': 0.219991} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.817854] env[63202]: DEBUG oslo_vmware.api [None req-ac13dc1f-36d9-4ead-8a3c-344061e61e29 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 987.817854] env[63202]: value = "task-1385693" [ 987.817854] env[63202]: _type = "Task" [ 987.817854] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.818083] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-eefe30e7-4320-41e9-881f-1b00135a970e tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 987.818251] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-eefe30e7-4320-41e9-881f-1b00135a970e tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 987.818752] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5e0cb08a-1660-4de7-b66e-370a08b36c03 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.830238] env[63202]: DEBUG oslo_vmware.api [None req-ac13dc1f-36d9-4ead-8a3c-344061e61e29 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385693, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.844348] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Acquiring lock "refresh_cache-8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.844503] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Acquired lock "refresh_cache-8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.844652] env[63202]: DEBUG nova.network.neutron [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 987.869867] env[63202]: DEBUG nova.policy [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2d292997e4c34051acc645b4a65dc369', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5e914f87450949b1a39866e8cfa3a0eb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 987.873734] env[63202]: DEBUG nova.network.neutron [req-eea299f2-babc-4ed7-9b3e-5739530a4e42 req-be703a76-fe50-4821-b24e-2235e681840c service nova] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Updated VIF entry in instance network info cache for port 0bd2810b-f01e-4cd8-9c3e-76ac2339047a. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 987.874102] env[63202]: DEBUG nova.network.neutron [req-eea299f2-babc-4ed7-9b3e-5739530a4e42 req-be703a76-fe50-4821-b24e-2235e681840c service nova] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Updating instance_info_cache with network_info: [{"id": "0bd2810b-f01e-4cd8-9c3e-76ac2339047a", "address": "fa:16:3e:9f:08:72", "network": {"id": "18390479-16ce-4012-9ba6-abf19b5616d1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1376026898-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22352c25bca8416a948014391a5389ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "085fb0ff-9285-4f1d-a008-a14da4844357", "external-id": "nsx-vlan-transportzone-729", "segmentation_id": 729, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0bd2810b-f0", "ovs_interfaceid": "0bd2810b-f01e-4cd8-9c3e-76ac2339047a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.893936] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-eefe30e7-4320-41e9-881f-1b00135a970e tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 987.894171] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-eefe30e7-4320-41e9-881f-1b00135a970e tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Deleting contents of the VM from datastore datastore2 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 987.894361] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-eefe30e7-4320-41e9-881f-1b00135a970e tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Deleting the datastore file [datastore2] f3128c09-3680-4b0b-b463-3d6cd203fcf4 {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 987.894634] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cb05b423-571f-4f1d-b720-1ced0f56ea04 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.901771] env[63202]: DEBUG oslo_vmware.api [None req-eefe30e7-4320-41e9-881f-1b00135a970e tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Waiting for the task: (returnval){ [ 987.901771] env[63202]: value = "task-1385695" [ 987.901771] env[63202]: _type = "Task" [ 987.901771] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.910421] env[63202]: DEBUG oslo_vmware.api [None req-eefe30e7-4320-41e9-881f-1b00135a970e tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Task: {'id': task-1385695, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.913556] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-95d4c183-16f0-4fd7-bdde-4f0d2160b43f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 987.913827] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6f238e00-c02d-4d63-bf73-e8506d32daad {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.921220] env[63202]: DEBUG oslo_vmware.api [None req-95d4c183-16f0-4fd7-bdde-4f0d2160b43f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Waiting for the task: (returnval){ [ 987.921220] env[63202]: value = "task-1385696" [ 987.921220] env[63202]: _type = "Task" [ 987.921220] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.929764] env[63202]: DEBUG oslo_vmware.api [None req-95d4c183-16f0-4fd7-bdde-4f0d2160b43f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385696, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.057119] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385689, 'name': CreateVM_Task} progress is 25%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.087244] env[63202]: DEBUG oslo_vmware.api [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385690, 'name': PowerOnVM_Task, 'duration_secs': 0.512297} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.087500] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 988.087705] env[63202]: INFO nova.compute.manager [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Took 7.31 seconds to spawn the instance on the hypervisor. [ 988.087884] env[63202]: DEBUG nova.compute.manager [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 988.088769] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cabe05d-d3b1-4ab8-a76e-7046f4b75395 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.152222] env[63202]: DEBUG nova.network.neutron [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Successfully created port: e9c43ed4-c64a-47ae-9574-2985bd75f032 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 988.307848] env[63202]: DEBUG nova.compute.manager [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 988.330495] env[63202]: DEBUG oslo_vmware.api [None req-ac13dc1f-36d9-4ead-8a3c-344061e61e29 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385693, 'name': ReconfigVM_Task, 'duration_secs': 0.17} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.330883] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-ac13dc1f-36d9-4ead-8a3c-344061e61e29 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294189', 'volume_id': 'd296fff5-0fd4-4820-8a58-c19b5b2d0da9', 'name': 'volume-d296fff5-0fd4-4820-8a58-c19b5b2d0da9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7437595c-fa35-483e-95f3-b75405b6bd13', 'attached_at': '', 'detached_at': '', 'volume_id': 'd296fff5-0fd4-4820-8a58-c19b5b2d0da9', 'serial': 'd296fff5-0fd4-4820-8a58-c19b5b2d0da9'} {{(pid=63202) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 988.381182] env[63202]: DEBUG oslo_concurrency.lockutils [req-eea299f2-babc-4ed7-9b3e-5739530a4e42 req-be703a76-fe50-4821-b24e-2235e681840c service nova] Releasing lock "refresh_cache-dd146c71-c391-41e8-8cc7-7276f4a518d9" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.389596] env[63202]: DEBUG nova.network.neutron [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 988.414404] env[63202]: DEBUG oslo_vmware.api [None req-eefe30e7-4320-41e9-881f-1b00135a970e tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Task: {'id': task-1385695, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.234907} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.418498] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-eefe30e7-4320-41e9-881f-1b00135a970e tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 988.418721] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-eefe30e7-4320-41e9-881f-1b00135a970e tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Deleted contents of the VM from datastore datastore2 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 988.418880] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-eefe30e7-4320-41e9-881f-1b00135a970e tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 988.419070] env[63202]: INFO nova.compute.manager [None req-eefe30e7-4320-41e9-881f-1b00135a970e tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Took 1.15 seconds to destroy the instance on the hypervisor. [ 988.419314] env[63202]: DEBUG oslo.service.loopingcall [None req-eefe30e7-4320-41e9-881f-1b00135a970e tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 988.419919] env[63202]: DEBUG nova.compute.manager [-] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 988.420013] env[63202]: DEBUG nova.network.neutron [-] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 988.430245] env[63202]: DEBUG oslo_vmware.api [None req-95d4c183-16f0-4fd7-bdde-4f0d2160b43f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385696, 'name': PowerOnVM_Task, 'duration_secs': 0.448361} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.432573] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-95d4c183-16f0-4fd7-bdde-4f0d2160b43f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 988.432775] env[63202]: DEBUG nova.compute.manager [None req-95d4c183-16f0-4fd7-bdde-4f0d2160b43f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 988.437576] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54ff5ccf-9453-4892-bcda-2f2c97010cc0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.538295] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e70a92c7-c0d5-4b46-995e-46c564fbc4de {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.552100] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b18f5b29-c836-43e2-b449-15e986880476 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.562310] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385689, 'name': CreateVM_Task, 'duration_secs': 1.194201} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.591257] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 988.592578] env[63202]: DEBUG nova.network.neutron [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Updating instance_info_cache with network_info: [{"id": "e7f10c86-d09d-4e51-a8f7-deef174f2539", "address": "fa:16:3e:8b:4a:de", "network": {"id": "69295efd-3c50-45e5-bc09-4f0cef6b8f0e", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1062402518-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f651d18a9559499da27f301ee75bb2fb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dced2f3d-7fd3-4a42-836d-9f02dab4c949", "external-id": "nsx-vlan-transportzone-117", "segmentation_id": 117, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7f10c86-d0", "ovs_interfaceid": "e7f10c86-d09d-4e51-a8f7-deef174f2539", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.594243] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.594420] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.594750] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 988.595508] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5123e71a-4068-4350-a154-a51ba898162e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.599502] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-32c295a5-27f1-42e5-b5a7-814aed540813 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.611854] env[63202]: INFO nova.compute.manager [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Took 17.00 seconds to build instance. [ 988.613670] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-142a2c7d-7a54-4e78-b120-812ef94e7618 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.617599] env[63202]: DEBUG oslo_vmware.api [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for the task: (returnval){ [ 988.617599] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52d6b55f-8a0a-cdf3-5bbb-504ed26122a9" [ 988.617599] env[63202]: _type = "Task" [ 988.617599] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.631884] env[63202]: DEBUG nova.compute.provider_tree [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 988.638389] env[63202]: DEBUG oslo_vmware.api [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52d6b55f-8a0a-cdf3-5bbb-504ed26122a9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.060498] env[63202]: DEBUG nova.compute.manager [req-cdcb4890-0cde-46f6-89a5-e7273e7359f4 req-603f8709-3d39-40ca-8bb6-4ce0fbc03f68 service nova] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Received event network-vif-deleted-be85354c-2bf7-4c06-898f-7061cdfada23 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 989.060498] env[63202]: INFO nova.compute.manager [req-cdcb4890-0cde-46f6-89a5-e7273e7359f4 req-603f8709-3d39-40ca-8bb6-4ce0fbc03f68 service nova] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Neutron deleted interface be85354c-2bf7-4c06-898f-7061cdfada23; detaching it from the instance and deleting it from the info cache [ 989.060498] env[63202]: DEBUG nova.network.neutron [req-cdcb4890-0cde-46f6-89a5-e7273e7359f4 req-603f8709-3d39-40ca-8bb6-4ce0fbc03f68 service nova] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.100535] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Releasing lock "refresh_cache-8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.100929] env[63202]: DEBUG nova.compute.manager [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Instance network_info: |[{"id": "e7f10c86-d09d-4e51-a8f7-deef174f2539", "address": "fa:16:3e:8b:4a:de", "network": {"id": "69295efd-3c50-45e5-bc09-4f0cef6b8f0e", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1062402518-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f651d18a9559499da27f301ee75bb2fb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dced2f3d-7fd3-4a42-836d-9f02dab4c949", "external-id": "nsx-vlan-transportzone-117", "segmentation_id": 117, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7f10c86-d0", "ovs_interfaceid": "e7f10c86-d09d-4e51-a8f7-deef174f2539", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 989.101392] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8b:4a:de', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dced2f3d-7fd3-4a42-836d-9f02dab4c949', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e7f10c86-d09d-4e51-a8f7-deef174f2539', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 989.109683] env[63202]: DEBUG oslo.service.loopingcall [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 989.110223] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 989.110556] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c5896b11-ff4b-4119-b930-33e6e6f551c4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.125394] env[63202]: DEBUG oslo_concurrency.lockutils [None req-d4f144e5-c010-4045-ad04-18ae3f2989ed tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "1bfb520e-6708-45a0-bb1f-2037fe1d801e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.519s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.135300] env[63202]: DEBUG oslo_vmware.api [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52d6b55f-8a0a-cdf3-5bbb-504ed26122a9, 'name': SearchDatastore_Task, 'duration_secs': 0.030467} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.136687] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.136958] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 989.137236] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.137370] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.137556] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 989.137811] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 989.137811] env[63202]: value = "task-1385697" [ 989.137811] env[63202]: _type = "Task" [ 989.137811] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.138050] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c40a8832-c411-4ac4-b84f-3ccb7a09bb41 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.140346] env[63202]: DEBUG nova.scheduler.client.report [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 989.153318] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385697, 'name': CreateVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.157898] env[63202]: DEBUG nova.compute.manager [req-83b621af-e90a-4f9d-a4d6-1f116da78143 req-301b2d8f-ce54-44c8-8e16-354f12b76ebb service nova] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Received event network-vif-plugged-e7f10c86-d09d-4e51-a8f7-deef174f2539 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 989.158123] env[63202]: DEBUG oslo_concurrency.lockutils [req-83b621af-e90a-4f9d-a4d6-1f116da78143 req-301b2d8f-ce54-44c8-8e16-354f12b76ebb service nova] Acquiring lock "8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.158262] env[63202]: DEBUG oslo_concurrency.lockutils [req-83b621af-e90a-4f9d-a4d6-1f116da78143 req-301b2d8f-ce54-44c8-8e16-354f12b76ebb service nova] Lock "8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.158404] env[63202]: DEBUG oslo_concurrency.lockutils [req-83b621af-e90a-4f9d-a4d6-1f116da78143 req-301b2d8f-ce54-44c8-8e16-354f12b76ebb service nova] Lock "8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.158577] env[63202]: DEBUG nova.compute.manager [req-83b621af-e90a-4f9d-a4d6-1f116da78143 req-301b2d8f-ce54-44c8-8e16-354f12b76ebb service nova] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] No waiting events found dispatching network-vif-plugged-e7f10c86-d09d-4e51-a8f7-deef174f2539 {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 989.158746] env[63202]: WARNING nova.compute.manager [req-83b621af-e90a-4f9d-a4d6-1f116da78143 req-301b2d8f-ce54-44c8-8e16-354f12b76ebb service nova] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Received unexpected event network-vif-plugged-e7f10c86-d09d-4e51-a8f7-deef174f2539 for instance with vm_state building and task_state spawning. [ 989.158936] env[63202]: DEBUG nova.compute.manager [req-83b621af-e90a-4f9d-a4d6-1f116da78143 req-301b2d8f-ce54-44c8-8e16-354f12b76ebb service nova] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Received event network-changed-e7f10c86-d09d-4e51-a8f7-deef174f2539 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 989.159068] env[63202]: DEBUG nova.compute.manager [req-83b621af-e90a-4f9d-a4d6-1f116da78143 req-301b2d8f-ce54-44c8-8e16-354f12b76ebb service nova] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Refreshing instance network info cache due to event network-changed-e7f10c86-d09d-4e51-a8f7-deef174f2539. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 989.159278] env[63202]: DEBUG oslo_concurrency.lockutils [req-83b621af-e90a-4f9d-a4d6-1f116da78143 req-301b2d8f-ce54-44c8-8e16-354f12b76ebb service nova] Acquiring lock "refresh_cache-8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.159462] env[63202]: DEBUG oslo_concurrency.lockutils [req-83b621af-e90a-4f9d-a4d6-1f116da78143 req-301b2d8f-ce54-44c8-8e16-354f12b76ebb service nova] Acquired lock "refresh_cache-8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.159654] env[63202]: DEBUG nova.network.neutron [req-83b621af-e90a-4f9d-a4d6-1f116da78143 req-301b2d8f-ce54-44c8-8e16-354f12b76ebb service nova] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Refreshing network info cache for port e7f10c86-d09d-4e51-a8f7-deef174f2539 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 989.206930] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 989.207157] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 989.208050] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7f075499-d8e1-4feb-99cf-6327c1c5b721 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.214680] env[63202]: DEBUG oslo_vmware.api [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for the task: (returnval){ [ 989.214680] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]526e6f9e-6126-b99b-2b64-85607d2e1626" [ 989.214680] env[63202]: _type = "Task" [ 989.214680] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.223822] env[63202]: DEBUG oslo_vmware.api [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]526e6f9e-6126-b99b-2b64-85607d2e1626, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.324308] env[63202]: DEBUG nova.compute.manager [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 989.349833] env[63202]: DEBUG nova.virt.hardware [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 989.350172] env[63202]: DEBUG nova.virt.hardware [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 989.350393] env[63202]: DEBUG nova.virt.hardware [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 989.350774] env[63202]: DEBUG nova.virt.hardware [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 989.350893] env[63202]: DEBUG nova.virt.hardware [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 989.351070] env[63202]: DEBUG nova.virt.hardware [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 989.351284] env[63202]: DEBUG nova.virt.hardware [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 989.351555] env[63202]: DEBUG nova.virt.hardware [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 989.351826] env[63202]: DEBUG nova.virt.hardware [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 989.352051] env[63202]: DEBUG nova.virt.hardware [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 989.352271] env[63202]: DEBUG nova.virt.hardware [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 989.353169] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90a46378-b235-4739-8543-ccd3401f613e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.362609] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d64448e3-dd2d-4d1c-85b8-5ddf7fd46eaa {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.367580] env[63202]: DEBUG nova.objects.instance [None req-ac13dc1f-36d9-4ead-8a3c-344061e61e29 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lazy-loading 'flavor' on Instance uuid 7437595c-fa35-483e-95f3-b75405b6bd13 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 989.539526] env[63202]: DEBUG nova.network.neutron [-] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.562947] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-67d3291c-9010-411b-80bc-f44881e90a9b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.580376] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66326cb9-699d-402d-99de-c1e6b9bac32b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.618888] env[63202]: DEBUG nova.compute.manager [req-cdcb4890-0cde-46f6-89a5-e7273e7359f4 req-603f8709-3d39-40ca-8bb6-4ce0fbc03f68 service nova] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Detach interface failed, port_id=be85354c-2bf7-4c06-898f-7061cdfada23, reason: Instance f3128c09-3680-4b0b-b463-3d6cd203fcf4 could not be found. {{(pid=63202) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 989.650721] env[63202]: DEBUG oslo_concurrency.lockutils [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.357s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.651136] env[63202]: DEBUG nova.compute.manager [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 989.654074] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385697, 'name': CreateVM_Task} progress is 99%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.655638] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8c9e3966-71f1-4bbb-b3ab-debaeebed328 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.083s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.655638] env[63202]: DEBUG nova.objects.instance [None req-8c9e3966-71f1-4bbb-b3ab-debaeebed328 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lazy-loading 'resources' on Instance uuid 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 989.726886] env[63202]: DEBUG oslo_vmware.api [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]526e6f9e-6126-b99b-2b64-85607d2e1626, 'name': SearchDatastore_Task, 'duration_secs': 0.054794} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.727820] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ecf74991-51ac-4b1d-b2dd-61ea6fd427be {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.734237] env[63202]: DEBUG oslo_vmware.api [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for the task: (returnval){ [ 989.734237] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]526c8243-f7c4-c82f-17ba-6c4dac5246a2" [ 989.734237] env[63202]: _type = "Task" [ 989.734237] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.743384] env[63202]: DEBUG oslo_vmware.api [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]526c8243-f7c4-c82f-17ba-6c4dac5246a2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.882394] env[63202]: DEBUG oslo_concurrency.lockutils [None req-ac13dc1f-36d9-4ead-8a3c-344061e61e29 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "7437595c-fa35-483e-95f3-b75405b6bd13" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.317s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.980359] env[63202]: DEBUG nova.network.neutron [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Successfully updated port: e9c43ed4-c64a-47ae-9574-2985bd75f032 {{(pid=63202) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 990.041288] env[63202]: INFO nova.compute.manager [-] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Took 1.62 seconds to deallocate network for instance. [ 990.099806] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "1bfb520e-6708-45a0-bb1f-2037fe1d801e" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.100202] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "1bfb520e-6708-45a0-bb1f-2037fe1d801e" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.100499] env[63202]: INFO nova.compute.manager [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Shelving [ 990.139327] env[63202]: DEBUG nova.network.neutron [req-83b621af-e90a-4f9d-a4d6-1f116da78143 req-301b2d8f-ce54-44c8-8e16-354f12b76ebb service nova] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Updated VIF entry in instance network info cache for port e7f10c86-d09d-4e51-a8f7-deef174f2539. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 990.139755] env[63202]: DEBUG nova.network.neutron [req-83b621af-e90a-4f9d-a4d6-1f116da78143 req-301b2d8f-ce54-44c8-8e16-354f12b76ebb service nova] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Updating instance_info_cache with network_info: [{"id": "e7f10c86-d09d-4e51-a8f7-deef174f2539", "address": "fa:16:3e:8b:4a:de", "network": {"id": "69295efd-3c50-45e5-bc09-4f0cef6b8f0e", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1062402518-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f651d18a9559499da27f301ee75bb2fb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dced2f3d-7fd3-4a42-836d-9f02dab4c949", "external-id": "nsx-vlan-transportzone-117", "segmentation_id": 117, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7f10c86-d0", "ovs_interfaceid": "e7f10c86-d09d-4e51-a8f7-deef174f2539", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.156690] env[63202]: DEBUG nova.compute.utils [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 990.170053] env[63202]: DEBUG nova.compute.manager [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 990.170053] env[63202]: DEBUG nova.network.neutron [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 990.173120] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385697, 'name': CreateVM_Task, 'duration_secs': 0.522646} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.174542] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 990.175741] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.175986] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.176434] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 990.176947] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-295b19bd-4d21-4bce-af23-85281e81670c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.182922] env[63202]: DEBUG oslo_vmware.api [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Waiting for the task: (returnval){ [ 990.182922] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52db5d2d-11f9-3302-b9ac-a1bd902edc87" [ 990.182922] env[63202]: _type = "Task" [ 990.182922] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.192661] env[63202]: DEBUG oslo_vmware.api [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52db5d2d-11f9-3302-b9ac-a1bd902edc87, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.225741] env[63202]: DEBUG nova.policy [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7dba72aed5d849ddb10117d7b797f0b3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '67083da1b3fc43bc91ed850937b771d7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 990.247782] env[63202]: DEBUG oslo_vmware.api [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]526c8243-f7c4-c82f-17ba-6c4dac5246a2, 'name': SearchDatastore_Task, 'duration_secs': 0.085003} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.248079] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.248341] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] dd146c71-c391-41e8-8cc7-7276f4a518d9/dd146c71-c391-41e8-8cc7-7276f4a518d9.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 990.248612] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5175811d-a282-4b20-b645-eddbc4e12cec {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.257834] env[63202]: DEBUG oslo_vmware.api [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for the task: (returnval){ [ 990.257834] env[63202]: value = "task-1385698" [ 990.257834] env[63202]: _type = "Task" [ 990.257834] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.272821] env[63202]: DEBUG oslo_vmware.api [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385698, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.373533] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ccb8a56-b2e5-41b5-8e60-cb5d160bc70d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.383217] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0721acb8-9197-4e32-ba0d-bbd9e7c697de {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.417352] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45a2b36a-be96-4045-b6cb-396567501138 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.426445] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d074b6cb-218b-4b69-b4f4-2d27b56ee19f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.441717] env[63202]: DEBUG nova.compute.provider_tree [None req-8c9e3966-71f1-4bbb-b3ab-debaeebed328 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 990.483357] env[63202]: DEBUG oslo_concurrency.lockutils [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquiring lock "refresh_cache-f23f2c04-c266-46a2-ad3f-34100f6246a0" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.483563] env[63202]: DEBUG oslo_concurrency.lockutils [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquired lock "refresh_cache-f23f2c04-c266-46a2-ad3f-34100f6246a0" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.483634] env[63202]: DEBUG nova.network.neutron [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 990.547265] env[63202]: DEBUG nova.network.neutron [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Successfully created port: 119b092b-6a1b-41ad-aa1c-b06e28b28f75 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 990.550235] env[63202]: DEBUG oslo_concurrency.lockutils [None req-eefe30e7-4320-41e9-881f-1b00135a970e tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.609103] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 990.609412] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-aaf80507-d919-48ea-a77a-6085dd686243 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.620912] env[63202]: DEBUG oslo_vmware.api [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 990.620912] env[63202]: value = "task-1385699" [ 990.620912] env[63202]: _type = "Task" [ 990.620912] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.640431] env[63202]: DEBUG oslo_vmware.api [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385699, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.652475] env[63202]: DEBUG oslo_concurrency.lockutils [req-83b621af-e90a-4f9d-a4d6-1f116da78143 req-301b2d8f-ce54-44c8-8e16-354f12b76ebb service nova] Releasing lock "refresh_cache-8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.674767] env[63202]: DEBUG nova.compute.manager [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 990.695170] env[63202]: DEBUG oslo_vmware.api [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52db5d2d-11f9-3302-b9ac-a1bd902edc87, 'name': SearchDatastore_Task, 'duration_secs': 0.01057} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.695577] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.695835] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 990.696136] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.696347] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.696521] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 990.696827] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f0975163-0f2a-481f-a92c-934db8b88b26 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.717850] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 990.717850] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 990.719451] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-316488fc-7ad4-4455-8699-de6da9121ba9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.730173] env[63202]: DEBUG oslo_vmware.api [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Waiting for the task: (returnval){ [ 990.730173] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52196943-bedc-01b9-e048-4e9444fb112b" [ 990.730173] env[63202]: _type = "Task" [ 990.730173] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.744762] env[63202]: DEBUG oslo_vmware.api [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52196943-bedc-01b9-e048-4e9444fb112b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.772117] env[63202]: DEBUG oslo_vmware.api [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385698, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.945527] env[63202]: DEBUG nova.scheduler.client.report [None req-8c9e3966-71f1-4bbb-b3ab-debaeebed328 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 991.016735] env[63202]: DEBUG nova.network.neutron [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 991.131608] env[63202]: DEBUG oslo_vmware.api [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385699, 'name': PowerOffVM_Task, 'duration_secs': 0.210284} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.131889] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 991.132914] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dbc4230-8f51-4264-904e-956f5a24773c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.152452] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14ab72d3-092f-4cc8-85ca-ec139968b22d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.157144] env[63202]: DEBUG nova.network.neutron [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Updating instance_info_cache with network_info: [{"id": "e9c43ed4-c64a-47ae-9574-2985bd75f032", "address": "fa:16:3e:74:d4:4a", "network": {"id": "e6c760f6-ded9-4905-b303-dbcde0d037b9", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1433567753-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e914f87450949b1a39866e8cfa3a0eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "26472e27-9835-4f87-ab7f-ca24dfee4e83", "external-id": "nsx-vlan-transportzone-335", "segmentation_id": 335, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape9c43ed4-c6", "ovs_interfaceid": "e9c43ed4-c64a-47ae-9574-2985bd75f032", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.181532] env[63202]: INFO nova.virt.block_device [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Booting with volume 6518fe7c-9ed9-49fb-b419-6c72d538c63d at /dev/sda [ 991.225968] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bd08f4b7-2968-48ab-ac3b-1c73c28ac99d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.240354] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6539362-1f9a-4003-8651-b86f2a116f28 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.261794] env[63202]: DEBUG nova.compute.manager [req-5d50a69a-4c34-4dc9-85bb-fd642af8521b req-fee6f13b-0c13-4231-b7d1-3d94f90c91d8 service nova] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Received event network-vif-plugged-e9c43ed4-c64a-47ae-9574-2985bd75f032 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 991.262122] env[63202]: DEBUG oslo_concurrency.lockutils [req-5d50a69a-4c34-4dc9-85bb-fd642af8521b req-fee6f13b-0c13-4231-b7d1-3d94f90c91d8 service nova] Acquiring lock "f23f2c04-c266-46a2-ad3f-34100f6246a0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.262436] env[63202]: DEBUG oslo_concurrency.lockutils [req-5d50a69a-4c34-4dc9-85bb-fd642af8521b req-fee6f13b-0c13-4231-b7d1-3d94f90c91d8 service nova] Lock "f23f2c04-c266-46a2-ad3f-34100f6246a0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.262889] env[63202]: DEBUG oslo_concurrency.lockutils [req-5d50a69a-4c34-4dc9-85bb-fd642af8521b req-fee6f13b-0c13-4231-b7d1-3d94f90c91d8 service nova] Lock "f23f2c04-c266-46a2-ad3f-34100f6246a0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.263017] env[63202]: DEBUG nova.compute.manager [req-5d50a69a-4c34-4dc9-85bb-fd642af8521b req-fee6f13b-0c13-4231-b7d1-3d94f90c91d8 service nova] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] No waiting events found dispatching network-vif-plugged-e9c43ed4-c64a-47ae-9574-2985bd75f032 {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 991.263373] env[63202]: WARNING nova.compute.manager [req-5d50a69a-4c34-4dc9-85bb-fd642af8521b req-fee6f13b-0c13-4231-b7d1-3d94f90c91d8 service nova] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Received unexpected event network-vif-plugged-e9c43ed4-c64a-47ae-9574-2985bd75f032 for instance with vm_state building and task_state spawning. [ 991.263432] env[63202]: DEBUG nova.compute.manager [req-5d50a69a-4c34-4dc9-85bb-fd642af8521b req-fee6f13b-0c13-4231-b7d1-3d94f90c91d8 service nova] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Received event network-changed-e9c43ed4-c64a-47ae-9574-2985bd75f032 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 991.263652] env[63202]: DEBUG nova.compute.manager [req-5d50a69a-4c34-4dc9-85bb-fd642af8521b req-fee6f13b-0c13-4231-b7d1-3d94f90c91d8 service nova] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Refreshing instance network info cache due to event network-changed-e9c43ed4-c64a-47ae-9574-2985bd75f032. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 991.263914] env[63202]: DEBUG oslo_concurrency.lockutils [req-5d50a69a-4c34-4dc9-85bb-fd642af8521b req-fee6f13b-0c13-4231-b7d1-3d94f90c91d8 service nova] Acquiring lock "refresh_cache-f23f2c04-c266-46a2-ad3f-34100f6246a0" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.264570] env[63202]: DEBUG oslo_vmware.api [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52196943-bedc-01b9-e048-4e9444fb112b, 'name': SearchDatastore_Task, 'duration_secs': 0.089002} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.269971] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-18f0b99a-18f5-45d3-ac28-a4726f13eb69 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.279761] env[63202]: DEBUG oslo_vmware.api [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385698, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.721647} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.290078] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] dd146c71-c391-41e8-8cc7-7276f4a518d9/dd146c71-c391-41e8-8cc7-7276f4a518d9.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 991.290395] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 991.291343] env[63202]: DEBUG oslo_vmware.api [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Waiting for the task: (returnval){ [ 991.291343] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]529efb97-6e74-49a5-523a-fc85ff07df97" [ 991.291343] env[63202]: _type = "Task" [ 991.291343] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.291555] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a507a9b6-59e4-47ff-bbef-a88a1cdc9d36 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.293528] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0734de1c-a8ca-4787-b32f-b54f42304da5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.307827] env[63202]: DEBUG oslo_vmware.api [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]529efb97-6e74-49a5-523a-fc85ff07df97, 'name': SearchDatastore_Task, 'duration_secs': 0.023687} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.309594] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.309964] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8/8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 991.312885] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-383eec90-dba5-4999-8e56-e4f252ab812c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.323610] env[63202]: DEBUG oslo_vmware.api [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for the task: (returnval){ [ 991.323610] env[63202]: value = "task-1385700" [ 991.323610] env[63202]: _type = "Task" [ 991.323610] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.323840] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-21739ce4-0383-4194-b002-d6ac2642d4bf {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.334610] env[63202]: DEBUG oslo_vmware.api [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385700, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.336153] env[63202]: DEBUG oslo_vmware.api [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Waiting for the task: (returnval){ [ 991.336153] env[63202]: value = "task-1385701" [ 991.336153] env[63202]: _type = "Task" [ 991.336153] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.348399] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc9356bc-aef8-4fe9-ad5a-f3e32313f68b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.353831] env[63202]: DEBUG oslo_vmware.api [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385701, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.358860] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08b3d372-64cd-4e51-8b60-c3a5a3d281b7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.377016] env[63202]: DEBUG nova.virt.block_device [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Updating existing volume attachment record: 22c996e0-c3ec-4821-91be-9d387553d14b {{(pid=63202) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 991.450943] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8c9e3966-71f1-4bbb-b3ab-debaeebed328 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.796s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.453254] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8b29a4d0-9d52-4d7e-8f06-7f965b6903bb tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.858s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.453486] env[63202]: DEBUG nova.objects.instance [None req-8b29a4d0-9d52-4d7e-8f06-7f965b6903bb tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lazy-loading 'resources' on Instance uuid eca51705-a972-48f6-85f5-6c397dad955d {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 991.467474] env[63202]: INFO nova.scheduler.client.report [None req-8c9e3966-71f1-4bbb-b3ab-debaeebed328 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Deleted allocations for instance 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e [ 991.665539] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Creating Snapshot of the VM instance {{(pid=63202) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 991.666118] env[63202]: DEBUG oslo_concurrency.lockutils [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Releasing lock "refresh_cache-f23f2c04-c266-46a2-ad3f-34100f6246a0" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.666535] env[63202]: DEBUG nova.compute.manager [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Instance network_info: |[{"id": "e9c43ed4-c64a-47ae-9574-2985bd75f032", "address": "fa:16:3e:74:d4:4a", "network": {"id": "e6c760f6-ded9-4905-b303-dbcde0d037b9", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1433567753-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e914f87450949b1a39866e8cfa3a0eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "26472e27-9835-4f87-ab7f-ca24dfee4e83", "external-id": "nsx-vlan-transportzone-335", "segmentation_id": 335, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape9c43ed4-c6", "ovs_interfaceid": "e9c43ed4-c64a-47ae-9574-2985bd75f032", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 991.667023] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-6c0a98e9-304f-4474-be45-a170c0136d80 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.669970] env[63202]: DEBUG oslo_concurrency.lockutils [req-5d50a69a-4c34-4dc9-85bb-fd642af8521b req-fee6f13b-0c13-4231-b7d1-3d94f90c91d8 service nova] Acquired lock "refresh_cache-f23f2c04-c266-46a2-ad3f-34100f6246a0" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.670218] env[63202]: DEBUG nova.network.neutron [req-5d50a69a-4c34-4dc9-85bb-fd642af8521b req-fee6f13b-0c13-4231-b7d1-3d94f90c91d8 service nova] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Refreshing network info cache for port e9c43ed4-c64a-47ae-9574-2985bd75f032 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 991.672033] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:74:d4:4a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '26472e27-9835-4f87-ab7f-ca24dfee4e83', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e9c43ed4-c64a-47ae-9574-2985bd75f032', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 991.683826] env[63202]: DEBUG oslo.service.loopingcall [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 991.685369] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 991.686344] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-63ae2bbd-d477-433b-9627-09539732455a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.713828] env[63202]: DEBUG oslo_vmware.api [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 991.713828] env[63202]: value = "task-1385702" [ 991.713828] env[63202]: _type = "Task" [ 991.713828] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.722049] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 991.722049] env[63202]: value = "task-1385703" [ 991.722049] env[63202]: _type = "Task" [ 991.722049] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.729527] env[63202]: DEBUG oslo_vmware.api [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385702, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.737103] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385703, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.810153] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquiring lock "3c6d286d-c152-46f1-b212-96a67324a56d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.810499] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "3c6d286d-c152-46f1-b212-96a67324a56d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.836424] env[63202]: DEBUG oslo_vmware.api [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385700, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065956} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.836802] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 991.841111] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4beb03e9-a1f5-42e0-a369-bb3858d68e31 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.850404] env[63202]: DEBUG oslo_vmware.api [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385701, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.868776] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Reconfiguring VM instance instance-00000058 to attach disk [datastore1] dd146c71-c391-41e8-8cc7-7276f4a518d9/dd146c71-c391-41e8-8cc7-7276f4a518d9.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 991.869429] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5cc7cad6-11fc-4c07-a958-5eb238bf78bc {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.891788] env[63202]: DEBUG oslo_vmware.api [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for the task: (returnval){ [ 991.891788] env[63202]: value = "task-1385704" [ 991.891788] env[63202]: _type = "Task" [ 991.891788] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.901300] env[63202]: DEBUG oslo_vmware.api [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385704, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.974496] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8c9e3966-71f1-4bbb-b3ab-debaeebed328 tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lock "7a72a1ab-3b3d-450d-a3a6-572b0035eb0e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.919s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.119791] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7012210e-21ca-46df-bd8f-15ecdd45fd4a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.130219] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eea6048-36fa-4f87-ba7d-8a624c61a994 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.165655] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d443e64f-c579-4c21-82cb-c0ed70e0d993 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.175331] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-121046c3-981b-4aa8-a831-8d8251d0cf3d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.194437] env[63202]: DEBUG nova.compute.provider_tree [None req-8b29a4d0-9d52-4d7e-8f06-7f965b6903bb tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 992.220994] env[63202]: DEBUG nova.compute.manager [req-041971fe-51e7-4271-836c-0ac57aea5746 req-1ae86f7b-11a7-4726-9cfb-113c9931d339 service nova] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Received event network-vif-plugged-119b092b-6a1b-41ad-aa1c-b06e28b28f75 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 992.221237] env[63202]: DEBUG oslo_concurrency.lockutils [req-041971fe-51e7-4271-836c-0ac57aea5746 req-1ae86f7b-11a7-4726-9cfb-113c9931d339 service nova] Acquiring lock "0d7fd297-37db-4876-bb91-5bb0b7dc335e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.221457] env[63202]: DEBUG oslo_concurrency.lockutils [req-041971fe-51e7-4271-836c-0ac57aea5746 req-1ae86f7b-11a7-4726-9cfb-113c9931d339 service nova] Lock "0d7fd297-37db-4876-bb91-5bb0b7dc335e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.221639] env[63202]: DEBUG oslo_concurrency.lockutils [req-041971fe-51e7-4271-836c-0ac57aea5746 req-1ae86f7b-11a7-4726-9cfb-113c9931d339 service nova] Lock "0d7fd297-37db-4876-bb91-5bb0b7dc335e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.221921] env[63202]: DEBUG nova.compute.manager [req-041971fe-51e7-4271-836c-0ac57aea5746 req-1ae86f7b-11a7-4726-9cfb-113c9931d339 service nova] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] No waiting events found dispatching network-vif-plugged-119b092b-6a1b-41ad-aa1c-b06e28b28f75 {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 992.222015] env[63202]: WARNING nova.compute.manager [req-041971fe-51e7-4271-836c-0ac57aea5746 req-1ae86f7b-11a7-4726-9cfb-113c9931d339 service nova] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Received unexpected event network-vif-plugged-119b092b-6a1b-41ad-aa1c-b06e28b28f75 for instance with vm_state building and task_state block_device_mapping. [ 992.229689] env[63202]: DEBUG oslo_vmware.api [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385702, 'name': CreateSnapshot_Task, 'duration_secs': 0.49937} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.232520] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Created Snapshot of the VM instance {{(pid=63202) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 992.233346] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e666a34-5c16-409f-b3f1-7c3ffb13b02d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.240323] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385703, 'name': CreateVM_Task} progress is 99%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.312666] env[63202]: DEBUG nova.compute.manager [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 992.347762] env[63202]: DEBUG oslo_vmware.api [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385701, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.597028} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.348149] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8/8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 992.348687] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 992.348885] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-383aee62-75e8-42f5-8ca4-0eb2dcfe8c29 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.358913] env[63202]: DEBUG oslo_vmware.api [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Waiting for the task: (returnval){ [ 992.358913] env[63202]: value = "task-1385705" [ 992.358913] env[63202]: _type = "Task" [ 992.358913] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.369133] env[63202]: DEBUG oslo_vmware.api [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385705, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.402124] env[63202]: DEBUG oslo_vmware.api [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385704, 'name': ReconfigVM_Task, 'duration_secs': 0.306002} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.404736] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Reconfigured VM instance instance-00000058 to attach disk [datastore1] dd146c71-c391-41e8-8cc7-7276f4a518d9/dd146c71-c391-41e8-8cc7-7276f4a518d9.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 992.406438] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f695bc5d-8876-439b-9805-1ca23d25d48a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.415854] env[63202]: DEBUG oslo_vmware.api [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for the task: (returnval){ [ 992.415854] env[63202]: value = "task-1385706" [ 992.415854] env[63202]: _type = "Task" [ 992.415854] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.430059] env[63202]: DEBUG oslo_vmware.api [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385706, 'name': Rename_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.448497] env[63202]: DEBUG nova.network.neutron [req-5d50a69a-4c34-4dc9-85bb-fd642af8521b req-fee6f13b-0c13-4231-b7d1-3d94f90c91d8 service nova] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Updated VIF entry in instance network info cache for port e9c43ed4-c64a-47ae-9574-2985bd75f032. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 992.448497] env[63202]: DEBUG nova.network.neutron [req-5d50a69a-4c34-4dc9-85bb-fd642af8521b req-fee6f13b-0c13-4231-b7d1-3d94f90c91d8 service nova] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Updating instance_info_cache with network_info: [{"id": "e9c43ed4-c64a-47ae-9574-2985bd75f032", "address": "fa:16:3e:74:d4:4a", "network": {"id": "e6c760f6-ded9-4905-b303-dbcde0d037b9", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1433567753-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e914f87450949b1a39866e8cfa3a0eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "26472e27-9835-4f87-ab7f-ca24dfee4e83", "external-id": "nsx-vlan-transportzone-335", "segmentation_id": 335, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape9c43ed4-c6", "ovs_interfaceid": "e9c43ed4-c64a-47ae-9574-2985bd75f032", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.457620] env[63202]: DEBUG nova.network.neutron [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Successfully updated port: 119b092b-6a1b-41ad-aa1c-b06e28b28f75 {{(pid=63202) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 992.701175] env[63202]: DEBUG nova.scheduler.client.report [None req-8b29a4d0-9d52-4d7e-8f06-7f965b6903bb tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 992.732643] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385703, 'name': CreateVM_Task, 'duration_secs': 0.577541} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.732806] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 992.733470] env[63202]: DEBUG oslo_concurrency.lockutils [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.733637] env[63202]: DEBUG oslo_concurrency.lockutils [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.733943] env[63202]: DEBUG oslo_concurrency.lockutils [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 992.734202] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-430f628e-873d-4996-99a3-7f0111abf292 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.739203] env[63202]: DEBUG oslo_vmware.api [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for the task: (returnval){ [ 992.739203] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52e00eef-9eb4-fc5f-1fbc-6506528c442a" [ 992.739203] env[63202]: _type = "Task" [ 992.739203] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.746785] env[63202]: DEBUG oslo_vmware.api [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52e00eef-9eb4-fc5f-1fbc-6506528c442a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.755846] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Creating linked-clone VM from snapshot {{(pid=63202) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 992.756260] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-2bc34aa8-0f9d-4e1d-ae97-bf56cd7e0e79 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.764700] env[63202]: DEBUG oslo_vmware.api [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 992.764700] env[63202]: value = "task-1385707" [ 992.764700] env[63202]: _type = "Task" [ 992.764700] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.773473] env[63202]: DEBUG oslo_vmware.api [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385707, 'name': CloneVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.836184] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.869149] env[63202]: DEBUG oslo_vmware.api [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385705, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07791} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.869422] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 992.870196] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a8b573f-db59-4537-b287-89428a102523 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.893065] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Reconfiguring VM instance instance-00000059 to attach disk [datastore1] 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8/8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 992.893272] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fe1654b2-62f0-4131-829f-ab639ba87e8c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.913457] env[63202]: DEBUG oslo_vmware.api [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Waiting for the task: (returnval){ [ 992.913457] env[63202]: value = "task-1385708" [ 992.913457] env[63202]: _type = "Task" [ 992.913457] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.929563] env[63202]: DEBUG oslo_vmware.api [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385706, 'name': Rename_Task, 'duration_secs': 0.15007} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.929796] env[63202]: DEBUG oslo_vmware.api [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385708, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.930085] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 992.930335] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a1fa8c7c-2e28-480c-8599-c38a5ea1d2cb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.937269] env[63202]: DEBUG oslo_vmware.api [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for the task: (returnval){ [ 992.937269] env[63202]: value = "task-1385709" [ 992.937269] env[63202]: _type = "Task" [ 992.937269] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.947654] env[63202]: DEBUG oslo_vmware.api [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385709, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.950327] env[63202]: DEBUG oslo_concurrency.lockutils [req-5d50a69a-4c34-4dc9-85bb-fd642af8521b req-fee6f13b-0c13-4231-b7d1-3d94f90c91d8 service nova] Releasing lock "refresh_cache-f23f2c04-c266-46a2-ad3f-34100f6246a0" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.963240] env[63202]: DEBUG oslo_concurrency.lockutils [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Acquiring lock "refresh_cache-0d7fd297-37db-4876-bb91-5bb0b7dc335e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.963398] env[63202]: DEBUG oslo_concurrency.lockutils [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Acquired lock "refresh_cache-0d7fd297-37db-4876-bb91-5bb0b7dc335e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.963515] env[63202]: DEBUG nova.network.neutron [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 993.206217] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8b29a4d0-9d52-4d7e-8f06-7f965b6903bb tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.753s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.210109] env[63202]: DEBUG oslo_concurrency.lockutils [None req-eefe30e7-4320-41e9-881f-1b00135a970e tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.659s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.210109] env[63202]: DEBUG nova.objects.instance [None req-eefe30e7-4320-41e9-881f-1b00135a970e tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Lazy-loading 'resources' on Instance uuid f3128c09-3680-4b0b-b463-3d6cd203fcf4 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 993.229200] env[63202]: INFO nova.scheduler.client.report [None req-8b29a4d0-9d52-4d7e-8f06-7f965b6903bb tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Deleted allocations for instance eca51705-a972-48f6-85f5-6c397dad955d [ 993.252695] env[63202]: DEBUG oslo_vmware.api [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52e00eef-9eb4-fc5f-1fbc-6506528c442a, 'name': SearchDatastore_Task, 'duration_secs': 0.009783} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.253110] env[63202]: DEBUG oslo_concurrency.lockutils [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.253378] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 993.253652] env[63202]: DEBUG oslo_concurrency.lockutils [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.253886] env[63202]: DEBUG oslo_concurrency.lockutils [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.254024] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 993.254313] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1c144858-adcd-450b-964d-1759815bf7b9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.270158] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 993.270419] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 993.271495] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dc3ba1a6-897c-4e76-a12b-f15d4673009b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.277095] env[63202]: DEBUG oslo_vmware.api [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385707, 'name': CloneVM_Task} progress is 94%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.282157] env[63202]: DEBUG oslo_vmware.api [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for the task: (returnval){ [ 993.282157] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52b3565e-205c-3d35-2e25-b2f5d8001d6f" [ 993.282157] env[63202]: _type = "Task" [ 993.282157] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.291563] env[63202]: DEBUG oslo_vmware.api [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52b3565e-205c-3d35-2e25-b2f5d8001d6f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.424268] env[63202]: DEBUG oslo_vmware.api [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385708, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.446979] env[63202]: DEBUG oslo_vmware.api [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385709, 'name': PowerOnVM_Task, 'duration_secs': 0.48103} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.447359] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 993.447487] env[63202]: INFO nova.compute.manager [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Took 8.67 seconds to spawn the instance on the hypervisor. [ 993.447636] env[63202]: DEBUG nova.compute.manager [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 993.448494] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fe15820-d187-465b-8130-2aa88db47fd7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.481419] env[63202]: DEBUG nova.compute.manager [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 993.483080] env[63202]: DEBUG nova.virt.hardware [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 993.483080] env[63202]: DEBUG nova.virt.hardware [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 993.483080] env[63202]: DEBUG nova.virt.hardware [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 993.483353] env[63202]: DEBUG nova.virt.hardware [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 993.483353] env[63202]: DEBUG nova.virt.hardware [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 993.483555] env[63202]: DEBUG nova.virt.hardware [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 993.483682] env[63202]: DEBUG nova.virt.hardware [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 993.484146] env[63202]: DEBUG nova.virt.hardware [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 993.484146] env[63202]: DEBUG nova.virt.hardware [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 993.484308] env[63202]: DEBUG nova.virt.hardware [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 993.484363] env[63202]: DEBUG nova.virt.hardware [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 993.485214] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a15c0793-e8ac-4a64-9908-cee38986b78f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.494871] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9250ced-a5e2-4325-87df-17efe3bbf1b1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.562365] env[63202]: DEBUG nova.network.neutron [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 993.740255] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8b29a4d0-9d52-4d7e-8f06-7f965b6903bb tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "eca51705-a972-48f6-85f5-6c397dad955d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.606s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.778599] env[63202]: DEBUG oslo_vmware.api [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385707, 'name': CloneVM_Task} progress is 94%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.796071] env[63202]: DEBUG oslo_vmware.api [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52b3565e-205c-3d35-2e25-b2f5d8001d6f, 'name': SearchDatastore_Task, 'duration_secs': 0.027651} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.803872] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4feb7476-04f1-4ac3-bea5-644c7e84e453 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.814141] env[63202]: DEBUG oslo_vmware.api [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for the task: (returnval){ [ 993.814141] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]521f3937-e48a-2336-0149-e6293a85c458" [ 993.814141] env[63202]: _type = "Task" [ 993.814141] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.826398] env[63202]: DEBUG oslo_vmware.api [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]521f3937-e48a-2336-0149-e6293a85c458, 'name': SearchDatastore_Task} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.826908] env[63202]: DEBUG nova.network.neutron [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Updating instance_info_cache with network_info: [{"id": "119b092b-6a1b-41ad-aa1c-b06e28b28f75", "address": "fa:16:3e:b3:0d:70", "network": {"id": "d956747e-1f39-460c-959e-e69ee83c040e", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-971941100-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "67083da1b3fc43bc91ed850937b771d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c66a277b-e3bf-43b8-a632-04fdd0720b91", "external-id": "nsx-vlan-transportzone-665", "segmentation_id": 665, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap119b092b-6a", "ovs_interfaceid": "119b092b-6a1b-41ad-aa1c-b06e28b28f75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.832027] env[63202]: DEBUG oslo_concurrency.lockutils [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.832027] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] f23f2c04-c266-46a2-ad3f-34100f6246a0/f23f2c04-c266-46a2-ad3f-34100f6246a0.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 993.832027] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ddad2a95-0e3f-48cd-a700-fe96fbad3c9b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.840333] env[63202]: DEBUG oslo_vmware.api [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for the task: (returnval){ [ 993.840333] env[63202]: value = "task-1385710" [ 993.840333] env[63202]: _type = "Task" [ 993.840333] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.853072] env[63202]: DEBUG oslo_vmware.api [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385710, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.897505] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51dfdb5b-f12d-4ef3-97b3-ad8be5bf74da {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.910690] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9c367bf-675c-4f90-9643-acb96557b882 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.931624] env[63202]: DEBUG oslo_vmware.api [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385708, 'name': ReconfigVM_Task, 'duration_secs': 0.832673} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.955170] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Reconfigured VM instance instance-00000059 to attach disk [datastore1] 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8/8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 993.956061] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-499a2503-a706-4e0d-9b13-8fdad15f34c9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.958356] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb6dbb27-eb9e-49e1-b4fd-52a4fec95263 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.968298] env[63202]: INFO nova.compute.manager [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Took 18.29 seconds to build instance. [ 993.976590] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3872e8b-d306-49b8-94eb-99cf4496474d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.979401] env[63202]: DEBUG oslo_vmware.api [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Waiting for the task: (returnval){ [ 993.979401] env[63202]: value = "task-1385711" [ 993.979401] env[63202]: _type = "Task" [ 993.979401] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.993132] env[63202]: DEBUG nova.compute.provider_tree [None req-eefe30e7-4320-41e9-881f-1b00135a970e tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 993.999567] env[63202]: DEBUG oslo_vmware.api [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385711, 'name': Rename_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.247480] env[63202]: DEBUG nova.compute.manager [req-e8c84916-e8a6-401d-8a57-564779fcfb4c req-370be3e9-f9c6-45de-87cc-3de4e5e94555 service nova] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Received event network-changed-119b092b-6a1b-41ad-aa1c-b06e28b28f75 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 994.247715] env[63202]: DEBUG nova.compute.manager [req-e8c84916-e8a6-401d-8a57-564779fcfb4c req-370be3e9-f9c6-45de-87cc-3de4e5e94555 service nova] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Refreshing instance network info cache due to event network-changed-119b092b-6a1b-41ad-aa1c-b06e28b28f75. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 994.247948] env[63202]: DEBUG oslo_concurrency.lockutils [req-e8c84916-e8a6-401d-8a57-564779fcfb4c req-370be3e9-f9c6-45de-87cc-3de4e5e94555 service nova] Acquiring lock "refresh_cache-0d7fd297-37db-4876-bb91-5bb0b7dc335e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.277811] env[63202]: DEBUG oslo_vmware.api [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385707, 'name': CloneVM_Task, 'duration_secs': 1.455698} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.278159] env[63202]: INFO nova.virt.vmwareapi.vmops [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Created linked-clone VM from snapshot [ 994.279175] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4bbee78-8510-44d9-9180-767cc5a84a06 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.291075] env[63202]: DEBUG nova.virt.vmwareapi.images [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Uploading image 294416a4-151a-4f4b-84f4-d0a689f9a708 {{(pid=63202) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 994.314895] env[63202]: DEBUG oslo_vmware.rw_handles [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 994.314895] env[63202]: value = "vm-294194" [ 994.314895] env[63202]: _type = "VirtualMachine" [ 994.314895] env[63202]: }. {{(pid=63202) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 994.315553] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-81bce1b3-64d1-45db-9d1f-b27e1813451e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.323540] env[63202]: DEBUG oslo_vmware.rw_handles [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lease: (returnval){ [ 994.323540] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52c22308-b526-6499-e8cb-434107149674" [ 994.323540] env[63202]: _type = "HttpNfcLease" [ 994.323540] env[63202]: } obtained for exporting VM: (result){ [ 994.323540] env[63202]: value = "vm-294194" [ 994.323540] env[63202]: _type = "VirtualMachine" [ 994.323540] env[63202]: }. {{(pid=63202) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 994.323853] env[63202]: DEBUG oslo_vmware.api [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the lease: (returnval){ [ 994.323853] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52c22308-b526-6499-e8cb-434107149674" [ 994.323853] env[63202]: _type = "HttpNfcLease" [ 994.323853] env[63202]: } to be ready. {{(pid=63202) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 994.330757] env[63202]: DEBUG oslo_concurrency.lockutils [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Releasing lock "refresh_cache-0d7fd297-37db-4876-bb91-5bb0b7dc335e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.331100] env[63202]: DEBUG nova.compute.manager [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Instance network_info: |[{"id": "119b092b-6a1b-41ad-aa1c-b06e28b28f75", "address": "fa:16:3e:b3:0d:70", "network": {"id": "d956747e-1f39-460c-959e-e69ee83c040e", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-971941100-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "67083da1b3fc43bc91ed850937b771d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c66a277b-e3bf-43b8-a632-04fdd0720b91", "external-id": "nsx-vlan-transportzone-665", "segmentation_id": 665, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap119b092b-6a", "ovs_interfaceid": "119b092b-6a1b-41ad-aa1c-b06e28b28f75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 994.331353] env[63202]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 994.331353] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52c22308-b526-6499-e8cb-434107149674" [ 994.331353] env[63202]: _type = "HttpNfcLease" [ 994.331353] env[63202]: } is initializing. {{(pid=63202) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 994.331592] env[63202]: DEBUG oslo_concurrency.lockutils [req-e8c84916-e8a6-401d-8a57-564779fcfb4c req-370be3e9-f9c6-45de-87cc-3de4e5e94555 service nova] Acquired lock "refresh_cache-0d7fd297-37db-4876-bb91-5bb0b7dc335e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.331774] env[63202]: DEBUG nova.network.neutron [req-e8c84916-e8a6-401d-8a57-564779fcfb4c req-370be3e9-f9c6-45de-87cc-3de4e5e94555 service nova] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Refreshing network info cache for port 119b092b-6a1b-41ad-aa1c-b06e28b28f75 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 994.332978] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b3:0d:70', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c66a277b-e3bf-43b8-a632-04fdd0720b91', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '119b092b-6a1b-41ad-aa1c-b06e28b28f75', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 994.340375] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Creating folder: Project (67083da1b3fc43bc91ed850937b771d7). Parent ref: group-v294090. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 994.343616] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a429b48a-428d-499d-997a-e561f7bb1df4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.356983] env[63202]: DEBUG oslo_vmware.api [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385710, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.496297} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.356983] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] f23f2c04-c266-46a2-ad3f-34100f6246a0/f23f2c04-c266-46a2-ad3f-34100f6246a0.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 994.357249] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 994.357613] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-669c57c8-d5ba-4b76-ac97-a0c110e96c98 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.361736] env[63202]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 994.361925] env[63202]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=63202) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 994.362346] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Folder already exists: Project (67083da1b3fc43bc91ed850937b771d7). Parent ref: group-v294090. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 994.362608] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Creating folder: Instances. Parent ref: group-v294179. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 994.363323] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-93ca0f60-2218-414d-a00a-74b89fbcf0c5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.367475] env[63202]: DEBUG oslo_vmware.api [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for the task: (returnval){ [ 994.367475] env[63202]: value = "task-1385714" [ 994.367475] env[63202]: _type = "Task" [ 994.367475] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.377342] env[63202]: DEBUG oslo_vmware.api [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385714, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.380670] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Created folder: Instances in parent group-v294179. [ 994.380891] env[63202]: DEBUG oslo.service.loopingcall [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 994.381479] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 994.381999] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9e099383-daa0-4249-a9d3-a5efe6125a14 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.402247] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 994.402247] env[63202]: value = "task-1385716" [ 994.402247] env[63202]: _type = "Task" [ 994.402247] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.410686] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385716, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.470349] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a258f676-a36c-418c-9ddc-5a5b4db0af30 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "dd146c71-c391-41e8-8cc7-7276f4a518d9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.802s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.493081] env[63202]: DEBUG oslo_vmware.api [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385711, 'name': Rename_Task, 'duration_secs': 0.200333} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.493744] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 994.493984] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ab9d27c3-9620-42a8-88a5-957665244cbd {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.496164] env[63202]: DEBUG nova.scheduler.client.report [None req-eefe30e7-4320-41e9-881f-1b00135a970e tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 994.503330] env[63202]: DEBUG oslo_concurrency.lockutils [None req-77680aa4-177c-4992-b6c2-c14dfc69934e tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "2110dfa7-1795-4c8f-92bf-18921409c99f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.503571] env[63202]: DEBUG oslo_concurrency.lockutils [None req-77680aa4-177c-4992-b6c2-c14dfc69934e tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "2110dfa7-1795-4c8f-92bf-18921409c99f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.503790] env[63202]: DEBUG oslo_concurrency.lockutils [None req-77680aa4-177c-4992-b6c2-c14dfc69934e tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "2110dfa7-1795-4c8f-92bf-18921409c99f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.504045] env[63202]: DEBUG oslo_concurrency.lockutils [None req-77680aa4-177c-4992-b6c2-c14dfc69934e tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "2110dfa7-1795-4c8f-92bf-18921409c99f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.504223] env[63202]: DEBUG oslo_concurrency.lockutils [None req-77680aa4-177c-4992-b6c2-c14dfc69934e tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "2110dfa7-1795-4c8f-92bf-18921409c99f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.507090] env[63202]: DEBUG oslo_vmware.api [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Waiting for the task: (returnval){ [ 994.507090] env[63202]: value = "task-1385717" [ 994.507090] env[63202]: _type = "Task" [ 994.507090] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.507549] env[63202]: INFO nova.compute.manager [None req-77680aa4-177c-4992-b6c2-c14dfc69934e tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Terminating instance [ 994.513015] env[63202]: DEBUG nova.compute.manager [None req-77680aa4-177c-4992-b6c2-c14dfc69934e tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 994.513264] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-77680aa4-177c-4992-b6c2-c14dfc69934e tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 994.514037] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faad4349-eef6-4d78-b11a-6e9439ccdacd {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.521015] env[63202]: DEBUG oslo_vmware.api [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385717, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.525948] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-77680aa4-177c-4992-b6c2-c14dfc69934e tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 994.526213] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3ef7c0ef-3511-4122-aa5b-dd41a574c03e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.532841] env[63202]: DEBUG oslo_vmware.api [None req-77680aa4-177c-4992-b6c2-c14dfc69934e tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 994.532841] env[63202]: value = "task-1385718" [ 994.532841] env[63202]: _type = "Task" [ 994.532841] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.542174] env[63202]: DEBUG oslo_vmware.api [None req-77680aa4-177c-4992-b6c2-c14dfc69934e tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385718, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.610784] env[63202]: DEBUG nova.network.neutron [req-e8c84916-e8a6-401d-8a57-564779fcfb4c req-370be3e9-f9c6-45de-87cc-3de4e5e94555 service nova] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Updated VIF entry in instance network info cache for port 119b092b-6a1b-41ad-aa1c-b06e28b28f75. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 994.611185] env[63202]: DEBUG nova.network.neutron [req-e8c84916-e8a6-401d-8a57-564779fcfb4c req-370be3e9-f9c6-45de-87cc-3de4e5e94555 service nova] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Updating instance_info_cache with network_info: [{"id": "119b092b-6a1b-41ad-aa1c-b06e28b28f75", "address": "fa:16:3e:b3:0d:70", "network": {"id": "d956747e-1f39-460c-959e-e69ee83c040e", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-971941100-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "67083da1b3fc43bc91ed850937b771d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c66a277b-e3bf-43b8-a632-04fdd0720b91", "external-id": "nsx-vlan-transportzone-665", "segmentation_id": 665, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap119b092b-6a", "ovs_interfaceid": "119b092b-6a1b-41ad-aa1c-b06e28b28f75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 994.832404] env[63202]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 994.832404] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52c22308-b526-6499-e8cb-434107149674" [ 994.832404] env[63202]: _type = "HttpNfcLease" [ 994.832404] env[63202]: } is ready. {{(pid=63202) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 994.833667] env[63202]: DEBUG oslo_vmware.rw_handles [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 994.833667] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52c22308-b526-6499-e8cb-434107149674" [ 994.833667] env[63202]: _type = "HttpNfcLease" [ 994.833667] env[63202]: }. {{(pid=63202) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 994.833667] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-456fa574-cb83-431e-9ee9-26547dbf8a13 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.841460] env[63202]: DEBUG oslo_vmware.rw_handles [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d3b529-2410-c84b-bc58-879df223fa53/disk-0.vmdk from lease info. {{(pid=63202) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 994.841650] env[63202]: DEBUG oslo_vmware.rw_handles [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d3b529-2410-c84b-bc58-879df223fa53/disk-0.vmdk for reading. {{(pid=63202) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 994.910514] env[63202]: DEBUG oslo_vmware.api [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385714, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063023} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.911214] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 994.912042] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb70b5e9-67a2-433a-81f6-10909a2f9e63 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.919767] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385716, 'name': CreateVM_Task, 'duration_secs': 0.404553} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.920156] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 994.920840] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'device_type': None, 'guest_format': None, 'mount_device': '/dev/sda', 'delete_on_termination': True, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294182', 'volume_id': '6518fe7c-9ed9-49fb-b419-6c72d538c63d', 'name': 'volume-6518fe7c-9ed9-49fb-b419-6c72d538c63d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0d7fd297-37db-4876-bb91-5bb0b7dc335e', 'attached_at': '', 'detached_at': '', 'volume_id': '6518fe7c-9ed9-49fb-b419-6c72d538c63d', 'serial': '6518fe7c-9ed9-49fb-b419-6c72d538c63d'}, 'disk_bus': None, 'boot_index': 0, 'attachment_id': '22c996e0-c3ec-4821-91be-9d387553d14b', 'volume_type': None}], 'swap': None} {{(pid=63202) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 994.921174] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Root volume attach. Driver type: vmdk {{(pid=63202) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 994.921855] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-505a7353-1b45-44a5-b11d-26c7a97ef2c3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.941535] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Reconfiguring VM instance instance-0000005a to attach disk [datastore1] f23f2c04-c266-46a2-ad3f-34100f6246a0/f23f2c04-c266-46a2-ad3f-34100f6246a0.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 994.942276] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bfca1716-2a17-4446-8bf3-26e13e387e88 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.966422] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4816f75f-4eca-49a4-ab14-e98626745064 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.969339] env[63202]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-0880b80d-4a34-450d-a12f-4544b1a68cfc {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.973211] env[63202]: DEBUG oslo_vmware.api [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for the task: (returnval){ [ 994.973211] env[63202]: value = "task-1385719" [ 994.973211] env[63202]: _type = "Task" [ 994.973211] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.981179] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f619bb6-2e24-42e8-a190-51b7dc020b33 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.987751] env[63202]: DEBUG oslo_vmware.api [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385719, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.000404] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-f20fcc3f-cb39-4806-b351-0dcabc448e37 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.003550] env[63202]: DEBUG oslo_concurrency.lockutils [None req-eefe30e7-4320-41e9-881f-1b00135a970e tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.794s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.006077] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.170s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.008293] env[63202]: INFO nova.compute.claims [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 995.022820] env[63202]: DEBUG oslo_vmware.api [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385717, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.023185] env[63202]: DEBUG oslo_vmware.api [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Waiting for the task: (returnval){ [ 995.023185] env[63202]: value = "task-1385720" [ 995.023185] env[63202]: _type = "Task" [ 995.023185] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.038026] env[63202]: DEBUG oslo_vmware.api [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Task: {'id': task-1385720, 'name': RelocateVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.046792] env[63202]: DEBUG oslo_vmware.api [None req-77680aa4-177c-4992-b6c2-c14dfc69934e tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385718, 'name': PowerOffVM_Task, 'duration_secs': 0.256184} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.047314] env[63202]: INFO nova.scheduler.client.report [None req-eefe30e7-4320-41e9-881f-1b00135a970e tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Deleted allocations for instance f3128c09-3680-4b0b-b463-3d6cd203fcf4 [ 995.048425] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-77680aa4-177c-4992-b6c2-c14dfc69934e tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 995.049093] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-77680aa4-177c-4992-b6c2-c14dfc69934e tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 995.049093] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-eec191b3-d1d1-4fd6-9bd5-4a80d52f0dd5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.114107] env[63202]: DEBUG oslo_concurrency.lockutils [req-e8c84916-e8a6-401d-8a57-564779fcfb4c req-370be3e9-f9c6-45de-87cc-3de4e5e94555 service nova] Releasing lock "refresh_cache-0d7fd297-37db-4876-bb91-5bb0b7dc335e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.143135] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-77680aa4-177c-4992-b6c2-c14dfc69934e tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 995.143478] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-77680aa4-177c-4992-b6c2-c14dfc69934e tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Deleting contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 995.143689] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-77680aa4-177c-4992-b6c2-c14dfc69934e tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Deleting the datastore file [datastore1] 2110dfa7-1795-4c8f-92bf-18921409c99f {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 995.144060] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-443a2d02-fbd4-436c-a381-463a6319c596 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.152350] env[63202]: DEBUG oslo_vmware.api [None req-77680aa4-177c-4992-b6c2-c14dfc69934e tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 995.152350] env[63202]: value = "task-1385722" [ 995.152350] env[63202]: _type = "Task" [ 995.152350] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.163387] env[63202]: DEBUG oslo_vmware.api [None req-77680aa4-177c-4992-b6c2-c14dfc69934e tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385722, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.484463] env[63202]: DEBUG oslo_vmware.api [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385719, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.523924] env[63202]: DEBUG oslo_vmware.api [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385717, 'name': PowerOnVM_Task, 'duration_secs': 0.524905} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.524753] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 995.524825] env[63202]: INFO nova.compute.manager [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Took 8.46 seconds to spawn the instance on the hypervisor. [ 995.524956] env[63202]: DEBUG nova.compute.manager [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 995.530016] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f05941f1-bead-456a-9c62-7bdfec52ef59 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.539475] env[63202]: DEBUG oslo_vmware.api [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Task: {'id': task-1385720, 'name': RelocateVM_Task} progress is 20%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.560145] env[63202]: DEBUG oslo_concurrency.lockutils [None req-eefe30e7-4320-41e9-881f-1b00135a970e tempest-ServersTestFqdnHostnames-1106038404 tempest-ServersTestFqdnHostnames-1106038404-project-member] Lock "f3128c09-3680-4b0b-b463-3d6cd203fcf4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.296s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.666338] env[63202]: DEBUG oslo_vmware.api [None req-77680aa4-177c-4992-b6c2-c14dfc69934e tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385722, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.276907} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.666867] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-77680aa4-177c-4992-b6c2-c14dfc69934e tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 995.667231] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-77680aa4-177c-4992-b6c2-c14dfc69934e tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Deleted contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 995.667508] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-77680aa4-177c-4992-b6c2-c14dfc69934e tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 995.667827] env[63202]: INFO nova.compute.manager [None req-77680aa4-177c-4992-b6c2-c14dfc69934e tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Took 1.15 seconds to destroy the instance on the hypervisor. [ 995.668242] env[63202]: DEBUG oslo.service.loopingcall [None req-77680aa4-177c-4992-b6c2-c14dfc69934e tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 995.668630] env[63202]: DEBUG nova.compute.manager [-] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 995.668769] env[63202]: DEBUG nova.network.neutron [-] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 995.984739] env[63202]: DEBUG oslo_vmware.api [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385719, 'name': ReconfigVM_Task, 'duration_secs': 0.598084} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.985125] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Reconfigured VM instance instance-0000005a to attach disk [datastore1] f23f2c04-c266-46a2-ad3f-34100f6246a0/f23f2c04-c266-46a2-ad3f-34100f6246a0.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 995.985971] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a5382eae-d63e-429f-8eef-8643c31bab06 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.995952] env[63202]: DEBUG oslo_vmware.api [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for the task: (returnval){ [ 995.995952] env[63202]: value = "task-1385723" [ 995.995952] env[63202]: _type = "Task" [ 995.995952] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.007032] env[63202]: DEBUG oslo_vmware.api [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385723, 'name': Rename_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.036853] env[63202]: DEBUG oslo_vmware.api [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Task: {'id': task-1385720, 'name': RelocateVM_Task, 'duration_secs': 0.71587} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.038470] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Volume attach. Driver type: vmdk {{(pid=63202) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 996.039015] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294182', 'volume_id': '6518fe7c-9ed9-49fb-b419-6c72d538c63d', 'name': 'volume-6518fe7c-9ed9-49fb-b419-6c72d538c63d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0d7fd297-37db-4876-bb91-5bb0b7dc335e', 'attached_at': '', 'detached_at': '', 'volume_id': '6518fe7c-9ed9-49fb-b419-6c72d538c63d', 'serial': '6518fe7c-9ed9-49fb-b419-6c72d538c63d'} {{(pid=63202) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 996.041099] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f022e656-51a9-4722-90bc-851338527bc4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.069111] env[63202]: INFO nova.compute.manager [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Took 15.78 seconds to build instance. [ 996.074817] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-693a59c3-bf29-4252-abc8-eb8b2c84b9bc {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.103364] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Reconfiguring VM instance instance-0000005b to attach disk [datastore2] volume-6518fe7c-9ed9-49fb-b419-6c72d538c63d/volume-6518fe7c-9ed9-49fb-b419-6c72d538c63d.vmdk or device None with type thin {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 996.107347] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8803b733-7f6a-4f47-9f9c-11049d824982 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.131758] env[63202]: DEBUG oslo_vmware.api [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Waiting for the task: (returnval){ [ 996.131758] env[63202]: value = "task-1385724" [ 996.131758] env[63202]: _type = "Task" [ 996.131758] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.145895] env[63202]: DEBUG oslo_vmware.api [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Task: {'id': task-1385724, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.255953] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82bf58f5-2250-4d0d-997f-cbcf9b2b7dd4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.265908] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71812c67-1780-472d-a46b-c6808d7f6f70 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.305574] env[63202]: DEBUG nova.compute.manager [req-0251cb96-deef-4fe7-ac92-a63cbeafcc97 req-7def6f16-b09d-47b9-ac78-f0ba6c3f74a3 service nova] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Received event network-vif-deleted-7d59550a-39cc-4a17-a5a8-a03665c6914a {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 996.305693] env[63202]: INFO nova.compute.manager [req-0251cb96-deef-4fe7-ac92-a63cbeafcc97 req-7def6f16-b09d-47b9-ac78-f0ba6c3f74a3 service nova] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Neutron deleted interface 7d59550a-39cc-4a17-a5a8-a03665c6914a; detaching it from the instance and deleting it from the info cache [ 996.305942] env[63202]: DEBUG nova.network.neutron [req-0251cb96-deef-4fe7-ac92-a63cbeafcc97 req-7def6f16-b09d-47b9-ac78-f0ba6c3f74a3 service nova] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.310242] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7a6a054-68d9-449c-97ff-2837592f9090 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.320661] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e349dd75-9014-4100-93a1-e552bf8243f9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.338223] env[63202]: DEBUG nova.compute.provider_tree [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 996.462464] env[63202]: DEBUG nova.network.neutron [-] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.508321] env[63202]: DEBUG oslo_vmware.api [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385723, 'name': Rename_Task, 'duration_secs': 0.193272} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.508922] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 996.509355] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5d78fdef-2d9d-4772-be14-2ce801bdf16e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.518166] env[63202]: DEBUG oslo_vmware.api [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for the task: (returnval){ [ 996.518166] env[63202]: value = "task-1385725" [ 996.518166] env[63202]: _type = "Task" [ 996.518166] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.528800] env[63202]: DEBUG oslo_vmware.api [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385725, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.532149] env[63202]: DEBUG nova.compute.manager [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Stashing vm_state: active {{(pid=63202) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 996.575086] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e8bbe7c8-8d35-4151-989e-a79ee38125da tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Lock "8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.320s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.641438] env[63202]: DEBUG oslo_vmware.api [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Task: {'id': task-1385724, 'name': ReconfigVM_Task, 'duration_secs': 0.440316} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.641846] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Reconfigured VM instance instance-0000005b to attach disk [datastore2] volume-6518fe7c-9ed9-49fb-b419-6c72d538c63d/volume-6518fe7c-9ed9-49fb-b419-6c72d538c63d.vmdk or device None with type thin {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 996.648429] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9ec2d20a-927e-4678-9e9c-39b2ad8273e9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.668315] env[63202]: DEBUG oslo_vmware.api [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Waiting for the task: (returnval){ [ 996.668315] env[63202]: value = "task-1385726" [ 996.668315] env[63202]: _type = "Task" [ 996.668315] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.680612] env[63202]: DEBUG oslo_vmware.api [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Task: {'id': task-1385726, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.815504] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-57cfc2f2-d297-4215-b6d0-aa3fdc865830 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.827026] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b10f4702-8e60-4e57-8307-9e7bb12a4516 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.843039] env[63202]: DEBUG nova.scheduler.client.report [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 996.860082] env[63202]: DEBUG nova.compute.manager [req-0251cb96-deef-4fe7-ac92-a63cbeafcc97 req-7def6f16-b09d-47b9-ac78-f0ba6c3f74a3 service nova] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Detach interface failed, port_id=7d59550a-39cc-4a17-a5a8-a03665c6914a, reason: Instance 2110dfa7-1795-4c8f-92bf-18921409c99f could not be found. {{(pid=63202) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 996.964744] env[63202]: INFO nova.compute.manager [-] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Took 1.30 seconds to deallocate network for instance. [ 997.045442] env[63202]: DEBUG oslo_vmware.api [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385725, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.097690] env[63202]: DEBUG oslo_concurrency.lockutils [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.184238] env[63202]: DEBUG oslo_vmware.api [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Task: {'id': task-1385726, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.348407] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.342s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.348671] env[63202]: DEBUG nova.compute.manager [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 997.351886] env[63202]: DEBUG oslo_concurrency.lockutils [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.255s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.471569] env[63202]: DEBUG oslo_concurrency.lockutils [None req-77680aa4-177c-4992-b6c2-c14dfc69934e tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.530094] env[63202]: DEBUG oslo_vmware.api [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385725, 'name': PowerOnVM_Task, 'duration_secs': 0.551005} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.530094] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 997.530094] env[63202]: INFO nova.compute.manager [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Took 8.21 seconds to spawn the instance on the hypervisor. [ 997.530255] env[63202]: DEBUG nova.compute.manager [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 997.531172] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-404fafe8-9178-4b1b-afbe-cd02af4d31f5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.670160] env[63202]: DEBUG nova.compute.manager [None req-7fd45976-c4d5-402b-94a8-8d5efafddff3 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 997.671400] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e054f6f-cb33-4799-a686-1e1ea863d223 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.688534] env[63202]: DEBUG oslo_vmware.api [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Task: {'id': task-1385726, 'name': ReconfigVM_Task, 'duration_secs': 1.003916} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.692037] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294182', 'volume_id': '6518fe7c-9ed9-49fb-b419-6c72d538c63d', 'name': 'volume-6518fe7c-9ed9-49fb-b419-6c72d538c63d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0d7fd297-37db-4876-bb91-5bb0b7dc335e', 'attached_at': '', 'detached_at': '', 'volume_id': '6518fe7c-9ed9-49fb-b419-6c72d538c63d', 'serial': '6518fe7c-9ed9-49fb-b419-6c72d538c63d'} {{(pid=63202) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 997.692037] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-27d24432-43a3-4776-bb2a-f3151a6c3e7e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.700487] env[63202]: DEBUG oslo_vmware.api [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Waiting for the task: (returnval){ [ 997.700487] env[63202]: value = "task-1385727" [ 997.700487] env[63202]: _type = "Task" [ 997.700487] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.711358] env[63202]: DEBUG oslo_vmware.api [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Task: {'id': task-1385727, 'name': Rename_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.855472] env[63202]: DEBUG nova.compute.utils [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 997.856894] env[63202]: DEBUG nova.compute.manager [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 997.857077] env[63202]: DEBUG nova.network.neutron [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 997.863334] env[63202]: INFO nova.compute.claims [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 997.920076] env[63202]: DEBUG nova.policy [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '63a29c785cf240a7b6418dadc7119574', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f7d28fea097541adadf0839940568409', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 998.054294] env[63202]: INFO nova.compute.manager [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Took 14.52 seconds to build instance. [ 998.193528] env[63202]: INFO nova.compute.manager [None req-7fd45976-c4d5-402b-94a8-8d5efafddff3 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] instance snapshotting [ 998.196324] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e21b4299-7120-42b0-9177-14816b6eb708 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.219408] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d978c099-50ce-47f1-9433-a414d5f48af1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.225516] env[63202]: DEBUG oslo_vmware.api [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Task: {'id': task-1385727, 'name': Rename_Task, 'duration_secs': 0.170925} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.226727] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 998.226998] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-17f83896-58d1-43f3-9fba-e75c41722b1d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.241030] env[63202]: DEBUG oslo_vmware.api [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Waiting for the task: (returnval){ [ 998.241030] env[63202]: value = "task-1385728" [ 998.241030] env[63202]: _type = "Task" [ 998.241030] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.253882] env[63202]: DEBUG oslo_vmware.api [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Task: {'id': task-1385728, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.368347] env[63202]: DEBUG nova.compute.manager [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 998.373385] env[63202]: INFO nova.compute.resource_tracker [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Updating resource usage from migration 839330ab-3902-4a6d-965a-18d61af14cb7 [ 998.557048] env[63202]: DEBUG oslo_concurrency.lockutils [None req-72616c8a-c6b0-483d-80a6-02eebbfe0c0f tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lock "f23f2c04-c266-46a2-ad3f-34100f6246a0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.034s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.568972] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28a4fe4b-0880-4e74-a906-dd9f2583cdc4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.580730] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a5c47e2-4c9c-46ba-b52e-4342ec89a639 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.587435] env[63202]: DEBUG nova.network.neutron [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Successfully created port: 131d0949-eb5d-4ee6-a244-d7898f8e7b0f {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 998.633464] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ed8e783-48bc-4096-a835-572d15ca8349 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.645120] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2d7152c-2f51-4b17-9f83-2c4e51bda1f1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.665865] env[63202]: DEBUG nova.compute.provider_tree [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 998.737495] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-7fd45976-c4d5-402b-94a8-8d5efafddff3 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Creating Snapshot of the VM instance {{(pid=63202) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 998.737495] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-af4600f8-bb26-4e92-a659-86cd746aa23b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.753163] env[63202]: DEBUG oslo_vmware.api [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Task: {'id': task-1385728, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.753910] env[63202]: DEBUG oslo_vmware.api [None req-7fd45976-c4d5-402b-94a8-8d5efafddff3 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Waiting for the task: (returnval){ [ 998.753910] env[63202]: value = "task-1385729" [ 998.753910] env[63202]: _type = "Task" [ 998.753910] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.763063] env[63202]: DEBUG oslo_vmware.api [None req-7fd45976-c4d5-402b-94a8-8d5efafddff3 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385729, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.169633] env[63202]: DEBUG nova.scheduler.client.report [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 999.253055] env[63202]: DEBUG oslo_vmware.api [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Task: {'id': task-1385728, 'name': PowerOnVM_Task, 'duration_secs': 0.74317} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.253377] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 999.253577] env[63202]: INFO nova.compute.manager [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Took 5.77 seconds to spawn the instance on the hypervisor. [ 999.253765] env[63202]: DEBUG nova.compute.manager [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 999.254579] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edd401df-a334-4da8-b278-e67764dde24c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.270381] env[63202]: DEBUG oslo_vmware.api [None req-7fd45976-c4d5-402b-94a8-8d5efafddff3 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385729, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.379297] env[63202]: DEBUG nova.compute.manager [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 999.407776] env[63202]: DEBUG nova.virt.hardware [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 999.408048] env[63202]: DEBUG nova.virt.hardware [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 999.408211] env[63202]: DEBUG nova.virt.hardware [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 999.408393] env[63202]: DEBUG nova.virt.hardware [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 999.408543] env[63202]: DEBUG nova.virt.hardware [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 999.408724] env[63202]: DEBUG nova.virt.hardware [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 999.408938] env[63202]: DEBUG nova.virt.hardware [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 999.409113] env[63202]: DEBUG nova.virt.hardware [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 999.409285] env[63202]: DEBUG nova.virt.hardware [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 999.409447] env[63202]: DEBUG nova.virt.hardware [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 999.409620] env[63202]: DEBUG nova.virt.hardware [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 999.410515] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b30c944e-3520-4de2-bd59-139933ab52a2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.419829] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc4d42e4-0a01-4093-ae9b-373be93c23bb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.675090] env[63202]: DEBUG oslo_concurrency.lockutils [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.323s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.675353] env[63202]: INFO nova.compute.manager [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Migrating [ 999.682184] env[63202]: DEBUG oslo_concurrency.lockutils [None req-77680aa4-177c-4992-b6c2-c14dfc69934e tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.211s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.682590] env[63202]: DEBUG nova.objects.instance [None req-77680aa4-177c-4992-b6c2-c14dfc69934e tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lazy-loading 'resources' on Instance uuid 2110dfa7-1795-4c8f-92bf-18921409c99f {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 999.767288] env[63202]: DEBUG oslo_vmware.api [None req-7fd45976-c4d5-402b-94a8-8d5efafddff3 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385729, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.780712] env[63202]: INFO nova.compute.manager [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Took 15.36 seconds to build instance. [ 999.818913] env[63202]: DEBUG nova.compute.manager [None req-f65e17c2-d94c-4c12-b5f8-52c3a981b23c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 999.819898] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bae4a883-36bd-4653-8fcf-6d63a2c4f3bb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.192571] env[63202]: DEBUG oslo_concurrency.lockutils [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "refresh_cache-dd146c71-c391-41e8-8cc7-7276f4a518d9" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.192974] env[63202]: DEBUG oslo_concurrency.lockutils [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquired lock "refresh_cache-dd146c71-c391-41e8-8cc7-7276f4a518d9" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.192974] env[63202]: DEBUG nova.network.neutron [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1000.272792] env[63202]: DEBUG oslo_vmware.api [None req-7fd45976-c4d5-402b-94a8-8d5efafddff3 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385729, 'name': CreateSnapshot_Task, 'duration_secs': 1.046855} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.272916] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-7fd45976-c4d5-402b-94a8-8d5efafddff3 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Created Snapshot of the VM instance {{(pid=63202) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1000.273778] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10ffcdbe-0d5f-4bd0-9619-40b4242c84ad {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.282460] env[63202]: DEBUG oslo_concurrency.lockutils [None req-44131156-d0b3-4691-acc5-5ea3a4a74661 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Lock "0d7fd297-37db-4876-bb91-5bb0b7dc335e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.867s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.335469] env[63202]: INFO nova.compute.manager [None req-f65e17c2-d94c-4c12-b5f8-52c3a981b23c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] instance snapshotting [ 1000.341433] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae7c2ae7-8c86-4e41-b22f-f0c83f596023 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.379021] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca77e898-3501-4641-a07f-57ac1d770ccc {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.396786] env[63202]: DEBUG nova.compute.manager [req-bad6946b-42e0-45ac-8891-9a173c0e529a req-07ebcecb-a1e1-446b-b56f-733a9c7141da service nova] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Received event network-changed-119b092b-6a1b-41ad-aa1c-b06e28b28f75 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1000.397345] env[63202]: DEBUG nova.compute.manager [req-bad6946b-42e0-45ac-8891-9a173c0e529a req-07ebcecb-a1e1-446b-b56f-733a9c7141da service nova] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Refreshing instance network info cache due to event network-changed-119b092b-6a1b-41ad-aa1c-b06e28b28f75. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1000.397708] env[63202]: DEBUG oslo_concurrency.lockutils [req-bad6946b-42e0-45ac-8891-9a173c0e529a req-07ebcecb-a1e1-446b-b56f-733a9c7141da service nova] Acquiring lock "refresh_cache-0d7fd297-37db-4876-bb91-5bb0b7dc335e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.397915] env[63202]: DEBUG oslo_concurrency.lockutils [req-bad6946b-42e0-45ac-8891-9a173c0e529a req-07ebcecb-a1e1-446b-b56f-733a9c7141da service nova] Acquired lock "refresh_cache-0d7fd297-37db-4876-bb91-5bb0b7dc335e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.398286] env[63202]: DEBUG nova.network.neutron [req-bad6946b-42e0-45ac-8891-9a173c0e529a req-07ebcecb-a1e1-446b-b56f-733a9c7141da service nova] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Refreshing network info cache for port 119b092b-6a1b-41ad-aa1c-b06e28b28f75 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1000.426858] env[63202]: DEBUG nova.compute.manager [req-8410b30a-a886-49d9-934b-cce314cf224d req-a953ef10-0883-4d36-b8d4-7852732d4b7e service nova] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Received event network-vif-plugged-131d0949-eb5d-4ee6-a244-d7898f8e7b0f {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1000.427055] env[63202]: DEBUG oslo_concurrency.lockutils [req-8410b30a-a886-49d9-934b-cce314cf224d req-a953ef10-0883-4d36-b8d4-7852732d4b7e service nova] Acquiring lock "3c6d286d-c152-46f1-b212-96a67324a56d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.427504] env[63202]: DEBUG oslo_concurrency.lockutils [req-8410b30a-a886-49d9-934b-cce314cf224d req-a953ef10-0883-4d36-b8d4-7852732d4b7e service nova] Lock "3c6d286d-c152-46f1-b212-96a67324a56d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.427743] env[63202]: DEBUG oslo_concurrency.lockutils [req-8410b30a-a886-49d9-934b-cce314cf224d req-a953ef10-0883-4d36-b8d4-7852732d4b7e service nova] Lock "3c6d286d-c152-46f1-b212-96a67324a56d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.427978] env[63202]: DEBUG nova.compute.manager [req-8410b30a-a886-49d9-934b-cce314cf224d req-a953ef10-0883-4d36-b8d4-7852732d4b7e service nova] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] No waiting events found dispatching network-vif-plugged-131d0949-eb5d-4ee6-a244-d7898f8e7b0f {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1000.428251] env[63202]: WARNING nova.compute.manager [req-8410b30a-a886-49d9-934b-cce314cf224d req-a953ef10-0883-4d36-b8d4-7852732d4b7e service nova] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Received unexpected event network-vif-plugged-131d0949-eb5d-4ee6-a244-d7898f8e7b0f for instance with vm_state building and task_state spawning. [ 1000.434373] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9207c8e0-4dec-490e-b20c-4d550124157f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.446743] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-430e48bf-5e95-4c02-b7ba-8d3fd6192c56 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.480483] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-256abd1e-e5bf-43d5-a0a9-14c418d73e5a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.489480] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e21d3663-06a6-44ba-acee-72c7cd45f925 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.509557] env[63202]: DEBUG nova.compute.provider_tree [None req-77680aa4-177c-4992-b6c2-c14dfc69934e tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1000.512275] env[63202]: DEBUG nova.network.neutron [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Successfully updated port: 131d0949-eb5d-4ee6-a244-d7898f8e7b0f {{(pid=63202) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1000.798434] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-7fd45976-c4d5-402b-94a8-8d5efafddff3 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Creating linked-clone VM from snapshot {{(pid=63202) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1000.798809] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-034ae4dd-d4ee-44b7-8afa-87c9988b599f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.810780] env[63202]: DEBUG oslo_vmware.api [None req-7fd45976-c4d5-402b-94a8-8d5efafddff3 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Waiting for the task: (returnval){ [ 1000.810780] env[63202]: value = "task-1385730" [ 1000.810780] env[63202]: _type = "Task" [ 1000.810780] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.820182] env[63202]: DEBUG oslo_vmware.api [None req-7fd45976-c4d5-402b-94a8-8d5efafddff3 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385730, 'name': CloneVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.895774] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-f65e17c2-d94c-4c12-b5f8-52c3a981b23c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Creating Snapshot of the VM instance {{(pid=63202) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1000.896162] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-ff75902f-83cb-4da0-84e5-d1b68be04e46 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.905422] env[63202]: DEBUG oslo_vmware.api [None req-f65e17c2-d94c-4c12-b5f8-52c3a981b23c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for the task: (returnval){ [ 1000.905422] env[63202]: value = "task-1385731" [ 1000.905422] env[63202]: _type = "Task" [ 1000.905422] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.915573] env[63202]: DEBUG oslo_vmware.api [None req-f65e17c2-d94c-4c12-b5f8-52c3a981b23c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385731, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.925680] env[63202]: DEBUG nova.network.neutron [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Updating instance_info_cache with network_info: [{"id": "0bd2810b-f01e-4cd8-9c3e-76ac2339047a", "address": "fa:16:3e:9f:08:72", "network": {"id": "18390479-16ce-4012-9ba6-abf19b5616d1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1376026898-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22352c25bca8416a948014391a5389ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "085fb0ff-9285-4f1d-a008-a14da4844357", "external-id": "nsx-vlan-transportzone-729", "segmentation_id": 729, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0bd2810b-f0", "ovs_interfaceid": "0bd2810b-f01e-4cd8-9c3e-76ac2339047a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.015332] env[63202]: DEBUG nova.scheduler.client.report [None req-77680aa4-177c-4992-b6c2-c14dfc69934e tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1001.020116] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquiring lock "refresh_cache-3c6d286d-c152-46f1-b212-96a67324a56d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.020116] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquired lock "refresh_cache-3c6d286d-c152-46f1-b212-96a67324a56d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.020116] env[63202]: DEBUG nova.network.neutron [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1001.157220] env[63202]: DEBUG nova.network.neutron [req-bad6946b-42e0-45ac-8891-9a173c0e529a req-07ebcecb-a1e1-446b-b56f-733a9c7141da service nova] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Updated VIF entry in instance network info cache for port 119b092b-6a1b-41ad-aa1c-b06e28b28f75. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1001.157706] env[63202]: DEBUG nova.network.neutron [req-bad6946b-42e0-45ac-8891-9a173c0e529a req-07ebcecb-a1e1-446b-b56f-733a9c7141da service nova] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Updating instance_info_cache with network_info: [{"id": "119b092b-6a1b-41ad-aa1c-b06e28b28f75", "address": "fa:16:3e:b3:0d:70", "network": {"id": "d956747e-1f39-460c-959e-e69ee83c040e", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-971941100-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.210", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "67083da1b3fc43bc91ed850937b771d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c66a277b-e3bf-43b8-a632-04fdd0720b91", "external-id": "nsx-vlan-transportzone-665", "segmentation_id": 665, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap119b092b-6a", "ovs_interfaceid": "119b092b-6a1b-41ad-aa1c-b06e28b28f75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.254636] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1001.254989] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1001.323418] env[63202]: DEBUG oslo_vmware.api [None req-7fd45976-c4d5-402b-94a8-8d5efafddff3 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385730, 'name': CloneVM_Task} progress is 94%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.417792] env[63202]: DEBUG oslo_vmware.api [None req-f65e17c2-d94c-4c12-b5f8-52c3a981b23c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385731, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.428930] env[63202]: DEBUG oslo_concurrency.lockutils [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Releasing lock "refresh_cache-dd146c71-c391-41e8-8cc7-7276f4a518d9" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.522914] env[63202]: DEBUG oslo_concurrency.lockutils [None req-77680aa4-177c-4992-b6c2-c14dfc69934e tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.841s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.546992] env[63202]: INFO nova.scheduler.client.report [None req-77680aa4-177c-4992-b6c2-c14dfc69934e tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Deleted allocations for instance 2110dfa7-1795-4c8f-92bf-18921409c99f [ 1001.555224] env[63202]: DEBUG nova.network.neutron [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1001.661086] env[63202]: DEBUG oslo_concurrency.lockutils [req-bad6946b-42e0-45ac-8891-9a173c0e529a req-07ebcecb-a1e1-446b-b56f-733a9c7141da service nova] Releasing lock "refresh_cache-0d7fd297-37db-4876-bb91-5bb0b7dc335e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.710484] env[63202]: DEBUG nova.network.neutron [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Updating instance_info_cache with network_info: [{"id": "131d0949-eb5d-4ee6-a244-d7898f8e7b0f", "address": "fa:16:3e:c8:01:19", "network": {"id": "0aa55fee-953e-4c2d-b88a-43b0cf86c7f3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-878071414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7d28fea097541adadf0839940568409", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap131d0949-eb", "ovs_interfaceid": "131d0949-eb5d-4ee6-a244-d7898f8e7b0f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.824352] env[63202]: DEBUG oslo_vmware.api [None req-7fd45976-c4d5-402b-94a8-8d5efafddff3 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385730, 'name': CloneVM_Task} progress is 94%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.918034] env[63202]: DEBUG oslo_vmware.api [None req-f65e17c2-d94c-4c12-b5f8-52c3a981b23c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385731, 'name': CreateSnapshot_Task, 'duration_secs': 0.568827} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.918306] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-f65e17c2-d94c-4c12-b5f8-52c3a981b23c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Created Snapshot of the VM instance {{(pid=63202) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1001.919175] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-411ac168-5e97-47a0-9abd-f6774a34c9e7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.055101] env[63202]: DEBUG oslo_concurrency.lockutils [None req-77680aa4-177c-4992-b6c2-c14dfc69934e tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "2110dfa7-1795-4c8f-92bf-18921409c99f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.551s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.213329] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Releasing lock "refresh_cache-3c6d286d-c152-46f1-b212-96a67324a56d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.213694] env[63202]: DEBUG nova.compute.manager [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Instance network_info: |[{"id": "131d0949-eb5d-4ee6-a244-d7898f8e7b0f", "address": "fa:16:3e:c8:01:19", "network": {"id": "0aa55fee-953e-4c2d-b88a-43b0cf86c7f3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-878071414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7d28fea097541adadf0839940568409", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap131d0949-eb", "ovs_interfaceid": "131d0949-eb5d-4ee6-a244-d7898f8e7b0f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1002.214177] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c8:01:19', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '131d0949-eb5d-4ee6-a244-d7898f8e7b0f', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1002.222482] env[63202]: DEBUG oslo.service.loopingcall [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1002.222726] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1002.223069] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-52e8d82e-6b5f-4298-aee8-9fd280511809 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.249600] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1002.249600] env[63202]: value = "task-1385732" [ 1002.249600] env[63202]: _type = "Task" [ 1002.249600] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.259945] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385732, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.323817] env[63202]: DEBUG oslo_vmware.api [None req-7fd45976-c4d5-402b-94a8-8d5efafddff3 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385730, 'name': CloneVM_Task} progress is 100%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.437319] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-f65e17c2-d94c-4c12-b5f8-52c3a981b23c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Creating linked-clone VM from snapshot {{(pid=63202) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1002.440305] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-72d5b411-0e9e-4df0-893b-fc278584c075 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.451425] env[63202]: DEBUG oslo_vmware.api [None req-f65e17c2-d94c-4c12-b5f8-52c3a981b23c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for the task: (returnval){ [ 1002.451425] env[63202]: value = "task-1385733" [ 1002.451425] env[63202]: _type = "Task" [ 1002.451425] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.458542] env[63202]: DEBUG nova.compute.manager [req-20fe60fa-fb9e-4d90-a9d1-1aad546c9ad6 req-141cbd13-b8a0-4ae1-b81a-52e3bb98a4de service nova] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Received event network-changed-131d0949-eb5d-4ee6-a244-d7898f8e7b0f {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1002.458842] env[63202]: DEBUG nova.compute.manager [req-20fe60fa-fb9e-4d90-a9d1-1aad546c9ad6 req-141cbd13-b8a0-4ae1-b81a-52e3bb98a4de service nova] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Refreshing instance network info cache due to event network-changed-131d0949-eb5d-4ee6-a244-d7898f8e7b0f. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1002.459198] env[63202]: DEBUG oslo_concurrency.lockutils [req-20fe60fa-fb9e-4d90-a9d1-1aad546c9ad6 req-141cbd13-b8a0-4ae1-b81a-52e3bb98a4de service nova] Acquiring lock "refresh_cache-3c6d286d-c152-46f1-b212-96a67324a56d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.459434] env[63202]: DEBUG oslo_concurrency.lockutils [req-20fe60fa-fb9e-4d90-a9d1-1aad546c9ad6 req-141cbd13-b8a0-4ae1-b81a-52e3bb98a4de service nova] Acquired lock "refresh_cache-3c6d286d-c152-46f1-b212-96a67324a56d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.459690] env[63202]: DEBUG nova.network.neutron [req-20fe60fa-fb9e-4d90-a9d1-1aad546c9ad6 req-141cbd13-b8a0-4ae1-b81a-52e3bb98a4de service nova] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Refreshing network info cache for port 131d0949-eb5d-4ee6-a244-d7898f8e7b0f {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1002.467646] env[63202]: DEBUG oslo_vmware.api [None req-f65e17c2-d94c-4c12-b5f8-52c3a981b23c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385733, 'name': CloneVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.760734] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385732, 'name': CreateVM_Task} progress is 99%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.825880] env[63202]: DEBUG oslo_vmware.api [None req-7fd45976-c4d5-402b-94a8-8d5efafddff3 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385730, 'name': CloneVM_Task, 'duration_secs': 1.60028} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.827816] env[63202]: INFO nova.virt.vmwareapi.vmops [None req-7fd45976-c4d5-402b-94a8-8d5efafddff3 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Created linked-clone VM from snapshot [ 1002.833082] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe5cc2b6-8703-4b32-b559-9a52f50409fe {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.837192] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "b29a8043-2006-4255-8036-669d2be594d3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.837570] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "b29a8043-2006-4255-8036-669d2be594d3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.846598] env[63202]: DEBUG nova.virt.vmwareapi.images [None req-7fd45976-c4d5-402b-94a8-8d5efafddff3 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Uploading image 857a011a-b565-47a2-af72-83e02755a311 {{(pid=63202) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1002.866717] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fd45976-c4d5-402b-94a8-8d5efafddff3 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Destroying the VM {{(pid=63202) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1002.867067] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-bc3342f2-39f3-4cdc-82ef-7266b1f1f543 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.876196] env[63202]: DEBUG oslo_vmware.api [None req-7fd45976-c4d5-402b-94a8-8d5efafddff3 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Waiting for the task: (returnval){ [ 1002.876196] env[63202]: value = "task-1385734" [ 1002.876196] env[63202]: _type = "Task" [ 1002.876196] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.886583] env[63202]: DEBUG oslo_vmware.api [None req-7fd45976-c4d5-402b-94a8-8d5efafddff3 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385734, 'name': Destroy_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.948540] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6627770c-c2c2-49f4-9134-96cfe9332346 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.971737] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Updating instance 'dd146c71-c391-41e8-8cc7-7276f4a518d9' progress to 0 {{(pid=63202) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1002.981303] env[63202]: DEBUG oslo_vmware.api [None req-f65e17c2-d94c-4c12-b5f8-52c3a981b23c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385733, 'name': CloneVM_Task} progress is 94%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.176322] env[63202]: DEBUG nova.network.neutron [req-20fe60fa-fb9e-4d90-a9d1-1aad546c9ad6 req-141cbd13-b8a0-4ae1-b81a-52e3bb98a4de service nova] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Updated VIF entry in instance network info cache for port 131d0949-eb5d-4ee6-a244-d7898f8e7b0f. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1003.176772] env[63202]: DEBUG nova.network.neutron [req-20fe60fa-fb9e-4d90-a9d1-1aad546c9ad6 req-141cbd13-b8a0-4ae1-b81a-52e3bb98a4de service nova] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Updating instance_info_cache with network_info: [{"id": "131d0949-eb5d-4ee6-a244-d7898f8e7b0f", "address": "fa:16:3e:c8:01:19", "network": {"id": "0aa55fee-953e-4c2d-b88a-43b0cf86c7f3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-878071414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7d28fea097541adadf0839940568409", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap131d0949-eb", "ovs_interfaceid": "131d0949-eb5d-4ee6-a244-d7898f8e7b0f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.255075] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1003.255319] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager.update_available_resource {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1003.262650] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385732, 'name': CreateVM_Task, 'duration_secs': 0.69584} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.263067] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1003.263516] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.263689] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.264027] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1003.264302] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-40c7a996-5025-40cc-bd7c-913cd2fc93e5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.270926] env[63202]: DEBUG oslo_vmware.api [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 1003.270926] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]525656d7-9c94-f27d-378d-0c17509cfc7a" [ 1003.270926] env[63202]: _type = "Task" [ 1003.270926] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.280959] env[63202]: DEBUG oslo_vmware.api [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]525656d7-9c94-f27d-378d-0c17509cfc7a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.341197] env[63202]: DEBUG nova.compute.manager [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: b29a8043-2006-4255-8036-669d2be594d3] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1003.389536] env[63202]: DEBUG oslo_vmware.api [None req-7fd45976-c4d5-402b-94a8-8d5efafddff3 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385734, 'name': Destroy_Task} progress is 33%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.464265] env[63202]: DEBUG oslo_vmware.api [None req-f65e17c2-d94c-4c12-b5f8-52c3a981b23c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385733, 'name': CloneVM_Task} progress is 94%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.479240] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1003.479548] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f0846435-a539-4483-9f11-e8bfad28a636 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.488239] env[63202]: DEBUG oslo_vmware.api [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for the task: (returnval){ [ 1003.488239] env[63202]: value = "task-1385735" [ 1003.488239] env[63202]: _type = "Task" [ 1003.488239] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.497774] env[63202]: DEBUG oslo_vmware.api [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385735, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.679955] env[63202]: DEBUG oslo_concurrency.lockutils [req-20fe60fa-fb9e-4d90-a9d1-1aad546c9ad6 req-141cbd13-b8a0-4ae1-b81a-52e3bb98a4de service nova] Releasing lock "refresh_cache-3c6d286d-c152-46f1-b212-96a67324a56d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.758477] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.758712] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.758881] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.759057] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63202) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1003.760098] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bf516f1-ff8a-4705-ba1c-bd6506503556 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.768970] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d587fc88-f196-47b7-bc9a-6a2603550382 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.789705] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5404043f-e400-4027-9579-97805aeb7ace {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.792327] env[63202]: DEBUG oslo_vmware.api [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]525656d7-9c94-f27d-378d-0c17509cfc7a, 'name': SearchDatastore_Task, 'duration_secs': 0.012574} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.792644] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.792873] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1003.793136] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.793298] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.793531] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1003.794152] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e59735ec-fb49-476c-bf74-e33fba1010de {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.799429] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2005cee5-b390-4b37-a705-6d9bbe1fb2bd {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.804588] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1003.804774] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1003.805897] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9eae7254-9ba2-45d0-9866-4bb9ae38ba56 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.835541] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179869MB free_disk=176GB free_vcpus=48 pci_devices=None {{(pid=63202) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1003.835757] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.835977] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.839054] env[63202]: DEBUG oslo_vmware.api [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 1003.839054] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52af0188-674d-dd1b-bd5e-b49f21a05a82" [ 1003.839054] env[63202]: _type = "Task" [ 1003.839054] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.855161] env[63202]: DEBUG oslo_vmware.api [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52af0188-674d-dd1b-bd5e-b49f21a05a82, 'name': SearchDatastore_Task, 'duration_secs': 0.010075} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.856258] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2530aef3-3d64-43e0-ad49-8d40138e0105 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.864832] env[63202]: DEBUG oslo_vmware.api [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 1003.864832] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]528d3038-05f6-a611-a1be-6bbb4ed25f83" [ 1003.864832] env[63202]: _type = "Task" [ 1003.864832] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.874361] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.878662] env[63202]: DEBUG oslo_vmware.api [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]528d3038-05f6-a611-a1be-6bbb4ed25f83, 'name': SearchDatastore_Task, 'duration_secs': 0.011272} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.882089] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.882367] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore2] 3c6d286d-c152-46f1-b212-96a67324a56d/3c6d286d-c152-46f1-b212-96a67324a56d.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1003.883691] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b5fce962-b14d-478b-877a-b260113ad33c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.892611] env[63202]: DEBUG oslo_vmware.api [None req-7fd45976-c4d5-402b-94a8-8d5efafddff3 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385734, 'name': Destroy_Task, 'duration_secs': 0.67669} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.893932] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-7fd45976-c4d5-402b-94a8-8d5efafddff3 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Destroyed the VM [ 1003.894205] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-7fd45976-c4d5-402b-94a8-8d5efafddff3 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Deleting Snapshot of the VM instance {{(pid=63202) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1003.894797] env[63202]: DEBUG oslo_vmware.api [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 1003.894797] env[63202]: value = "task-1385736" [ 1003.894797] env[63202]: _type = "Task" [ 1003.894797] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.895030] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-73061c3f-6fa0-4c29-ba1e-3ca3fd9c1068 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.908965] env[63202]: DEBUG oslo_vmware.api [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385736, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.910015] env[63202]: DEBUG oslo_vmware.api [None req-7fd45976-c4d5-402b-94a8-8d5efafddff3 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Waiting for the task: (returnval){ [ 1003.910015] env[63202]: value = "task-1385737" [ 1003.910015] env[63202]: _type = "Task" [ 1003.910015] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.919359] env[63202]: DEBUG oslo_vmware.api [None req-7fd45976-c4d5-402b-94a8-8d5efafddff3 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385737, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.963830] env[63202]: DEBUG oslo_vmware.api [None req-f65e17c2-d94c-4c12-b5f8-52c3a981b23c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385733, 'name': CloneVM_Task} progress is 100%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.998939] env[63202]: DEBUG oslo_vmware.api [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385735, 'name': PowerOffVM_Task, 'duration_secs': 0.300079} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.999307] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1003.999544] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Updating instance 'dd146c71-c391-41e8-8cc7-7276f4a518d9' progress to 17 {{(pid=63202) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1004.406873] env[63202]: DEBUG oslo_vmware.api [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385736, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.48148} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.407207] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore2] 3c6d286d-c152-46f1-b212-96a67324a56d/3c6d286d-c152-46f1-b212-96a67324a56d.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1004.407484] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1004.407765] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f0a8f84d-4e67-4471-8ccc-3d29928f0447 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.416500] env[63202]: DEBUG oslo_vmware.api [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 1004.416500] env[63202]: value = "task-1385738" [ 1004.416500] env[63202]: _type = "Task" [ 1004.416500] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.422914] env[63202]: DEBUG oslo_vmware.api [None req-7fd45976-c4d5-402b-94a8-8d5efafddff3 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385737, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.428302] env[63202]: DEBUG oslo_vmware.api [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385738, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.465384] env[63202]: DEBUG oslo_vmware.api [None req-f65e17c2-d94c-4c12-b5f8-52c3a981b23c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385733, 'name': CloneVM_Task, 'duration_secs': 1.527685} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.465592] env[63202]: INFO nova.virt.vmwareapi.vmops [None req-f65e17c2-d94c-4c12-b5f8-52c3a981b23c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Created linked-clone VM from snapshot [ 1004.466421] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3284fd4b-65d6-4118-8c4b-3b2795158f50 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.474625] env[63202]: DEBUG nova.virt.vmwareapi.images [None req-f65e17c2-d94c-4c12-b5f8-52c3a981b23c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Uploading image 39830181-3a83-45ed-9dae-2cd9e605a46a {{(pid=63202) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1004.487880] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-f65e17c2-d94c-4c12-b5f8-52c3a981b23c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Destroying the VM {{(pid=63202) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1004.488193] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-1ac17036-c728-4f84-ac97-b7b6892db45c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.496423] env[63202]: DEBUG oslo_vmware.api [None req-f65e17c2-d94c-4c12-b5f8-52c3a981b23c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for the task: (returnval){ [ 1004.496423] env[63202]: value = "task-1385739" [ 1004.496423] env[63202]: _type = "Task" [ 1004.496423] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.507913] env[63202]: DEBUG nova.virt.hardware [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1004.507913] env[63202]: DEBUG nova.virt.hardware [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1004.508132] env[63202]: DEBUG nova.virt.hardware [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1004.508281] env[63202]: DEBUG nova.virt.hardware [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1004.508434] env[63202]: DEBUG nova.virt.hardware [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1004.508588] env[63202]: DEBUG nova.virt.hardware [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1004.508802] env[63202]: DEBUG nova.virt.hardware [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1004.508955] env[63202]: DEBUG nova.virt.hardware [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1004.509147] env[63202]: DEBUG nova.virt.hardware [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1004.509354] env[63202]: DEBUG nova.virt.hardware [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1004.509477] env[63202]: DEBUG nova.virt.hardware [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1004.514692] env[63202]: DEBUG oslo_vmware.api [None req-f65e17c2-d94c-4c12-b5f8-52c3a981b23c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385739, 'name': Destroy_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.514955] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d47dfa7b-bb5b-4c08-8a59-132766066450 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.539030] env[63202]: DEBUG oslo_vmware.api [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for the task: (returnval){ [ 1004.539030] env[63202]: value = "task-1385740" [ 1004.539030] env[63202]: _type = "Task" [ 1004.539030] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.851293] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Applying migration context for instance dd146c71-c391-41e8-8cc7-7276f4a518d9 as it has an incoming, in-progress migration 839330ab-3902-4a6d-965a-18d61af14cb7. Migration status is migrating {{(pid=63202) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1004.852714] env[63202]: INFO nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Updating resource usage from migration 839330ab-3902-4a6d-965a-18d61af14cb7 [ 1004.871502] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1004.871652] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 7437595c-fa35-483e-95f3-b75405b6bd13 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1004.871841] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 8437f856-a707-49c5-b8eb-5a22cdb990f8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1004.871951] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 1bfb520e-6708-45a0-bb1f-2037fe1d801e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1004.872119] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1004.872252] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance f23f2c04-c266-46a2-ad3f-34100f6246a0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1004.872374] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 0d7fd297-37db-4876-bb91-5bb0b7dc335e actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1004.872490] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 3c6d286d-c152-46f1-b212-96a67324a56d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1004.872645] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Migration 839330ab-3902-4a6d-965a-18d61af14cb7 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1004.872867] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance dd146c71-c391-41e8-8cc7-7276f4a518d9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1004.924747] env[63202]: DEBUG oslo_vmware.api [None req-7fd45976-c4d5-402b-94a8-8d5efafddff3 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385737, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.930822] env[63202]: DEBUG oslo_vmware.api [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385738, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.121576} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.931102] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1004.931885] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed5b0856-9ac4-4e46-8a7e-35c77b3a464d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.953979] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] 3c6d286d-c152-46f1-b212-96a67324a56d/3c6d286d-c152-46f1-b212-96a67324a56d.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1004.956030] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0f8abf5c-91ae-4b05-b120-f4f4320800f3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.976222] env[63202]: DEBUG oslo_vmware.rw_handles [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d3b529-2410-c84b-bc58-879df223fa53/disk-0.vmdk. {{(pid=63202) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1004.977097] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b54f6836-a13b-4d2e-9e93-854536077106 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.980667] env[63202]: DEBUG oslo_vmware.api [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 1004.980667] env[63202]: value = "task-1385741" [ 1004.980667] env[63202]: _type = "Task" [ 1004.980667] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.985463] env[63202]: DEBUG oslo_vmware.rw_handles [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d3b529-2410-c84b-bc58-879df223fa53/disk-0.vmdk is in state: ready. {{(pid=63202) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1004.985629] env[63202]: ERROR oslo_vmware.rw_handles [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d3b529-2410-c84b-bc58-879df223fa53/disk-0.vmdk due to incomplete transfer. [ 1004.986193] env[63202]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-ea286ff2-eded-4f84-87d3-c590fd2bf7f9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.990843] env[63202]: DEBUG oslo_vmware.api [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385741, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.994498] env[63202]: DEBUG oslo_vmware.rw_handles [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d3b529-2410-c84b-bc58-879df223fa53/disk-0.vmdk. {{(pid=63202) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1004.994692] env[63202]: DEBUG nova.virt.vmwareapi.images [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Uploaded image 294416a4-151a-4f4b-84f4-d0a689f9a708 to the Glance image server {{(pid=63202) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1004.997156] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Destroying the VM {{(pid=63202) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1004.997489] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-bdcecc02-427a-4e3b-b3cc-ffdfbf9c0e25 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.007645] env[63202]: DEBUG oslo_vmware.api [None req-f65e17c2-d94c-4c12-b5f8-52c3a981b23c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385739, 'name': Destroy_Task} progress is 33%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.008924] env[63202]: DEBUG oslo_vmware.api [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 1005.008924] env[63202]: value = "task-1385742" [ 1005.008924] env[63202]: _type = "Task" [ 1005.008924] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.017254] env[63202]: DEBUG oslo_vmware.api [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385742, 'name': Destroy_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.049420] env[63202]: DEBUG oslo_vmware.api [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385740, 'name': ReconfigVM_Task, 'duration_secs': 0.311848} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.049710] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Updating instance 'dd146c71-c391-41e8-8cc7-7276f4a518d9' progress to 33 {{(pid=63202) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1005.377078] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance b29a8043-2006-4255-8036-669d2be594d3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1005.377408] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Total usable vcpus: 48, total allocated vcpus: 10 {{(pid=63202) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1005.377562] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2496MB phys_disk=200GB used_disk=9GB total_vcpus=48 used_vcpus=10 pci_stats=[] {{(pid=63202) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1005.424183] env[63202]: DEBUG oslo_vmware.api [None req-7fd45976-c4d5-402b-94a8-8d5efafddff3 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385737, 'name': RemoveSnapshot_Task, 'duration_secs': 1.233017} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.426838] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-7fd45976-c4d5-402b-94a8-8d5efafddff3 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Deleted Snapshot of the VM instance {{(pid=63202) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1005.493156] env[63202]: DEBUG oslo_vmware.api [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385741, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.510596] env[63202]: DEBUG oslo_vmware.api [None req-f65e17c2-d94c-4c12-b5f8-52c3a981b23c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385739, 'name': Destroy_Task, 'duration_secs': 0.623459} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.513645] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-f65e17c2-d94c-4c12-b5f8-52c3a981b23c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Destroyed the VM [ 1005.513903] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-f65e17c2-d94c-4c12-b5f8-52c3a981b23c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Deleting Snapshot of the VM instance {{(pid=63202) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1005.516404] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-fdc41452-d1cd-4b95-bb9e-43c33d86a2cb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.524940] env[63202]: DEBUG oslo_vmware.api [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385742, 'name': Destroy_Task} progress is 33%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.526406] env[63202]: DEBUG oslo_vmware.api [None req-f65e17c2-d94c-4c12-b5f8-52c3a981b23c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for the task: (returnval){ [ 1005.526406] env[63202]: value = "task-1385743" [ 1005.526406] env[63202]: _type = "Task" [ 1005.526406] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.539115] env[63202]: DEBUG oslo_vmware.api [None req-f65e17c2-d94c-4c12-b5f8-52c3a981b23c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385743, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.544483] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe49a998-75ff-46e3-b370-4e0ec2cf703b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.553538] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c45fe815-65f1-4195-ac45-dd3ba5106893 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.558688] env[63202]: DEBUG nova.virt.hardware [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1005.558957] env[63202]: DEBUG nova.virt.hardware [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1005.559146] env[63202]: DEBUG nova.virt.hardware [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1005.559328] env[63202]: DEBUG nova.virt.hardware [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1005.559522] env[63202]: DEBUG nova.virt.hardware [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1005.559696] env[63202]: DEBUG nova.virt.hardware [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1005.559924] env[63202]: DEBUG nova.virt.hardware [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1005.560149] env[63202]: DEBUG nova.virt.hardware [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1005.560340] env[63202]: DEBUG nova.virt.hardware [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1005.560526] env[63202]: DEBUG nova.virt.hardware [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1005.560740] env[63202]: DEBUG nova.virt.hardware [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1005.566104] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Reconfiguring VM instance instance-00000058 to detach disk 2000 {{(pid=63202) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1005.566393] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-38551c67-9062-4606-9033-c8992dd076d7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.610391] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff3606a8-1d11-4022-ab14-107027ee926d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.613397] env[63202]: DEBUG oslo_vmware.api [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for the task: (returnval){ [ 1005.613397] env[63202]: value = "task-1385744" [ 1005.613397] env[63202]: _type = "Task" [ 1005.613397] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.632386] env[63202]: DEBUG oslo_vmware.api [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385744, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.636297] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aea50ba1-5f10-48a9-b061-a8c179e43144 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.654015] env[63202]: DEBUG nova.compute.provider_tree [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Updating inventory in ProviderTree for provider 79b33d17-3e75-494c-a550-67b275de2079 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1005.931304] env[63202]: WARNING nova.compute.manager [None req-7fd45976-c4d5-402b-94a8-8d5efafddff3 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Image not found during snapshot: nova.exception.ImageNotFound: Image 857a011a-b565-47a2-af72-83e02755a311 could not be found. [ 1005.992298] env[63202]: DEBUG oslo_vmware.api [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385741, 'name': ReconfigVM_Task, 'duration_secs': 0.574553} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.992553] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Reconfigured VM instance instance-0000005c to attach disk [datastore2] 3c6d286d-c152-46f1-b212-96a67324a56d/3c6d286d-c152-46f1-b212-96a67324a56d.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1005.993221] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e98bb5a0-9ffe-45e4-a8cb-e6fc3f1e0a30 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.000341] env[63202]: DEBUG oslo_vmware.api [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 1006.000341] env[63202]: value = "task-1385745" [ 1006.000341] env[63202]: _type = "Task" [ 1006.000341] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.010141] env[63202]: DEBUG oslo_vmware.api [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385745, 'name': Rename_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.019650] env[63202]: DEBUG oslo_vmware.api [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385742, 'name': Destroy_Task, 'duration_secs': 0.957616} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.019895] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Destroyed the VM [ 1006.020154] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Deleting Snapshot of the VM instance {{(pid=63202) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1006.020396] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-2fa805bc-7282-411b-a8c0-87090842d5bc {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.026519] env[63202]: DEBUG oslo_vmware.api [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 1006.026519] env[63202]: value = "task-1385746" [ 1006.026519] env[63202]: _type = "Task" [ 1006.026519] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.037980] env[63202]: DEBUG oslo_vmware.api [None req-f65e17c2-d94c-4c12-b5f8-52c3a981b23c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385743, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.040842] env[63202]: DEBUG oslo_vmware.api [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385746, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.122674] env[63202]: DEBUG oslo_vmware.api [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385744, 'name': ReconfigVM_Task, 'duration_secs': 0.205272} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.122912] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Reconfigured VM instance instance-00000058 to detach disk 2000 {{(pid=63202) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1006.123706] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07701159-ecb3-481c-8a27-ca4eeaa43f0f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.146075] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Reconfiguring VM instance instance-00000058 to attach disk [datastore1] dd146c71-c391-41e8-8cc7-7276f4a518d9/dd146c71-c391-41e8-8cc7-7276f4a518d9.vmdk or device None with type thin {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1006.146358] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5b9a6021-9931-4ee4-898b-2bf1882d7b06 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.167136] env[63202]: DEBUG oslo_vmware.api [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for the task: (returnval){ [ 1006.167136] env[63202]: value = "task-1385747" [ 1006.167136] env[63202]: _type = "Task" [ 1006.167136] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.175500] env[63202]: DEBUG oslo_vmware.api [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385747, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.177756] env[63202]: ERROR nova.scheduler.client.report [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [req-783dbead-0621-4c60-a11d-81679b75564e] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 79b33d17-3e75-494c-a550-67b275de2079. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-783dbead-0621-4c60-a11d-81679b75564e"}]} [ 1006.194951] env[63202]: DEBUG nova.scheduler.client.report [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Refreshing inventories for resource provider 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1006.209234] env[63202]: DEBUG nova.scheduler.client.report [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Updating ProviderTree inventory for provider 79b33d17-3e75-494c-a550-67b275de2079 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1006.209437] env[63202]: DEBUG nova.compute.provider_tree [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Updating inventory in ProviderTree for provider 79b33d17-3e75-494c-a550-67b275de2079 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1006.220680] env[63202]: DEBUG nova.scheduler.client.report [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Refreshing aggregate associations for resource provider 79b33d17-3e75-494c-a550-67b275de2079, aggregates: None {{(pid=63202) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1006.241396] env[63202]: DEBUG nova.scheduler.client.report [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Refreshing trait associations for resource provider 79b33d17-3e75-494c-a550-67b275de2079, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,HW_ARCH_X86_64,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO {{(pid=63202) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1006.382310] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ec529b6-3b66-4cee-aa8c-7048e7a5ffc0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.391140] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d07da0a-70de-4817-8e41-ed3efca997fe {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.425914] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fddca77-4149-4773-afb1-2dff2bf48fee {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.433706] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-267c33a7-7782-4559-84fe-c935427a3a76 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.448845] env[63202]: DEBUG nova.compute.provider_tree [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Updating inventory in ProviderTree for provider 79b33d17-3e75-494c-a550-67b275de2079 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1006.511412] env[63202]: DEBUG oslo_vmware.api [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385745, 'name': Rename_Task, 'duration_secs': 0.204219} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.511688] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1006.511947] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5af923ba-7d3e-4ba7-8e9f-c4474fdb939e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.520525] env[63202]: DEBUG oslo_vmware.api [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 1006.520525] env[63202]: value = "task-1385748" [ 1006.520525] env[63202]: _type = "Task" [ 1006.520525] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.537054] env[63202]: DEBUG oslo_vmware.api [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385748, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.542555] env[63202]: DEBUG oslo_vmware.api [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385746, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.545481] env[63202]: DEBUG oslo_vmware.api [None req-f65e17c2-d94c-4c12-b5f8-52c3a981b23c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385743, 'name': RemoveSnapshot_Task, 'duration_secs': 0.606976} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.545767] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-f65e17c2-d94c-4c12-b5f8-52c3a981b23c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Deleted Snapshot of the VM instance {{(pid=63202) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1006.679241] env[63202]: DEBUG oslo_vmware.api [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385747, 'name': ReconfigVM_Task, 'duration_secs': 0.271696} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.679685] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Reconfigured VM instance instance-00000058 to attach disk [datastore1] dd146c71-c391-41e8-8cc7-7276f4a518d9/dd146c71-c391-41e8-8cc7-7276f4a518d9.vmdk or device None with type thin {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1006.680088] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Updating instance 'dd146c71-c391-41e8-8cc7-7276f4a518d9' progress to 50 {{(pid=63202) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1007.010211] env[63202]: DEBUG nova.scheduler.client.report [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Updated inventory for provider 79b33d17-3e75-494c-a550-67b275de2079 with generation 111 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1007.010211] env[63202]: DEBUG nova.compute.provider_tree [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Updating resource provider 79b33d17-3e75-494c-a550-67b275de2079 generation from 111 to 112 during operation: update_inventory {{(pid=63202) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1007.010211] env[63202]: DEBUG nova.compute.provider_tree [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Updating inventory in ProviderTree for provider 79b33d17-3e75-494c-a550-67b275de2079 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1007.034335] env[63202]: DEBUG oslo_vmware.api [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385748, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.039920] env[63202]: DEBUG oslo_vmware.api [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385746, 'name': RemoveSnapshot_Task, 'duration_secs': 0.717349} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.040248] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Deleted Snapshot of the VM instance {{(pid=63202) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1007.040516] env[63202]: DEBUG nova.compute.manager [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1007.041325] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ebc3033-39e1-4580-b96c-cdfac5d89db6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.051744] env[63202]: WARNING nova.compute.manager [None req-f65e17c2-d94c-4c12-b5f8-52c3a981b23c tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Image not found during snapshot: nova.exception.ImageNotFound: Image 39830181-3a83-45ed-9dae-2cd9e605a46a could not be found. [ 1007.190156] env[63202]: DEBUG oslo_concurrency.lockutils [None req-937a425e-6dad-4384-9f38-0330ccbf0c18 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Acquiring lock "8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.190505] env[63202]: DEBUG oslo_concurrency.lockutils [None req-937a425e-6dad-4384-9f38-0330ccbf0c18 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Lock "8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.190794] env[63202]: DEBUG oslo_concurrency.lockutils [None req-937a425e-6dad-4384-9f38-0330ccbf0c18 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Acquiring lock "8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.191056] env[63202]: DEBUG oslo_concurrency.lockutils [None req-937a425e-6dad-4384-9f38-0330ccbf0c18 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Lock "8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.191294] env[63202]: DEBUG oslo_concurrency.lockutils [None req-937a425e-6dad-4384-9f38-0330ccbf0c18 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Lock "8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.193575] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eb7db38-b8c8-46dd-8620-94bdc008a205 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.196545] env[63202]: INFO nova.compute.manager [None req-937a425e-6dad-4384-9f38-0330ccbf0c18 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Terminating instance [ 1007.198462] env[63202]: DEBUG nova.compute.manager [None req-937a425e-6dad-4384-9f38-0330ccbf0c18 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1007.198750] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-937a425e-6dad-4384-9f38-0330ccbf0c18 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1007.200014] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-380b884a-8ba1-4e32-8973-19d2d3bd2c11 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.217559] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e40a9ad6-c720-4bf7-b9f3-8a4d4c182991 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.222399] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-937a425e-6dad-4384-9f38-0330ccbf0c18 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1007.223083] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ec18c866-26d2-495e-bd3b-d1e341218c7f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.242087] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Updating instance 'dd146c71-c391-41e8-8cc7-7276f4a518d9' progress to 67 {{(pid=63202) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1007.248061] env[63202]: DEBUG oslo_vmware.api [None req-937a425e-6dad-4384-9f38-0330ccbf0c18 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Waiting for the task: (returnval){ [ 1007.248061] env[63202]: value = "task-1385749" [ 1007.248061] env[63202]: _type = "Task" [ 1007.248061] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.257056] env[63202]: DEBUG oslo_vmware.api [None req-937a425e-6dad-4384-9f38-0330ccbf0c18 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385749, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.495626] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63202) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1007.496077] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.660s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.496189] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.622s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.497698] env[63202]: INFO nova.compute.claims [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: b29a8043-2006-4255-8036-669d2be594d3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1007.531123] env[63202]: DEBUG oslo_vmware.api [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385748, 'name': PowerOnVM_Task, 'duration_secs': 0.832236} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.531412] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1007.531595] env[63202]: INFO nova.compute.manager [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Took 8.15 seconds to spawn the instance on the hypervisor. [ 1007.531773] env[63202]: DEBUG nova.compute.manager [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1007.532571] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab16aac5-e747-49eb-897d-1f92dadda16c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.555398] env[63202]: INFO nova.compute.manager [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Shelve offloading [ 1007.556882] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1007.557127] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c198908b-c6ab-4601-85e7-2b0d1375a04e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.564330] env[63202]: DEBUG oslo_vmware.api [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 1007.564330] env[63202]: value = "task-1385750" [ 1007.564330] env[63202]: _type = "Task" [ 1007.564330] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.574743] env[63202]: DEBUG oslo_vmware.api [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385750, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.621966] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1f61d9fe-563d-45cf-80c0-eb3b6899f2fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquiring lock "f23f2c04-c266-46a2-ad3f-34100f6246a0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.622275] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1f61d9fe-563d-45cf-80c0-eb3b6899f2fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lock "f23f2c04-c266-46a2-ad3f-34100f6246a0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.622490] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1f61d9fe-563d-45cf-80c0-eb3b6899f2fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquiring lock "f23f2c04-c266-46a2-ad3f-34100f6246a0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.622707] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1f61d9fe-563d-45cf-80c0-eb3b6899f2fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lock "f23f2c04-c266-46a2-ad3f-34100f6246a0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.622896] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1f61d9fe-563d-45cf-80c0-eb3b6899f2fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lock "f23f2c04-c266-46a2-ad3f-34100f6246a0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.625079] env[63202]: INFO nova.compute.manager [None req-1f61d9fe-563d-45cf-80c0-eb3b6899f2fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Terminating instance [ 1007.627314] env[63202]: DEBUG nova.compute.manager [None req-1f61d9fe-563d-45cf-80c0-eb3b6899f2fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1007.627517] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-1f61d9fe-563d-45cf-80c0-eb3b6899f2fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1007.628394] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26ee478d-b9b6-4076-8dec-babd8f0d7a08 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.638631] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f61d9fe-563d-45cf-80c0-eb3b6899f2fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1007.638876] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cb04cd6f-e3e9-4f2b-bf0f-49631f0dcc7b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.646208] env[63202]: DEBUG oslo_vmware.api [None req-1f61d9fe-563d-45cf-80c0-eb3b6899f2fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for the task: (returnval){ [ 1007.646208] env[63202]: value = "task-1385751" [ 1007.646208] env[63202]: _type = "Task" [ 1007.646208] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.654902] env[63202]: DEBUG oslo_vmware.api [None req-1f61d9fe-563d-45cf-80c0-eb3b6899f2fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385751, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.762039] env[63202]: DEBUG oslo_vmware.api [None req-937a425e-6dad-4384-9f38-0330ccbf0c18 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385749, 'name': PowerOffVM_Task, 'duration_secs': 0.226358} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.762306] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-937a425e-6dad-4384-9f38-0330ccbf0c18 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1007.762418] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-937a425e-6dad-4384-9f38-0330ccbf0c18 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1007.762659] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f352c347-72c6-41b8-977f-9a2498a791f5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.790283] env[63202]: DEBUG nova.network.neutron [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Port 0bd2810b-f01e-4cd8-9c3e-76ac2339047a binding to destination host cpu-1 is already ACTIVE {{(pid=63202) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1007.896711] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-937a425e-6dad-4384-9f38-0330ccbf0c18 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1007.896711] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-937a425e-6dad-4384-9f38-0330ccbf0c18 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Deleting contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1007.896711] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-937a425e-6dad-4384-9f38-0330ccbf0c18 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Deleting the datastore file [datastore1] 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8 {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1007.897063] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a9fef42c-2d4c-42dc-97ab-8ebd52614c04 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.905326] env[63202]: DEBUG oslo_vmware.api [None req-937a425e-6dad-4384-9f38-0330ccbf0c18 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Waiting for the task: (returnval){ [ 1007.905326] env[63202]: value = "task-1385753" [ 1007.905326] env[63202]: _type = "Task" [ 1007.905326] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.913747] env[63202]: DEBUG oslo_vmware.api [None req-937a425e-6dad-4384-9f38-0330ccbf0c18 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385753, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.051203] env[63202]: INFO nova.compute.manager [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Took 15.23 seconds to build instance. [ 1008.080488] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] VM already powered off {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1008.080861] env[63202]: DEBUG nova.compute.manager [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1008.082116] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3c372b1-b6f2-423c-b1b5-50050b018a52 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.091491] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "refresh_cache-1bfb520e-6708-45a0-bb1f-2037fe1d801e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.091784] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquired lock "refresh_cache-1bfb520e-6708-45a0-bb1f-2037fe1d801e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.092112] env[63202]: DEBUG nova.network.neutron [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1008.156668] env[63202]: DEBUG oslo_vmware.api [None req-1f61d9fe-563d-45cf-80c0-eb3b6899f2fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385751, 'name': PowerOffVM_Task, 'duration_secs': 0.387102} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.156930] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f61d9fe-563d-45cf-80c0-eb3b6899f2fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1008.157096] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-1f61d9fe-563d-45cf-80c0-eb3b6899f2fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1008.157358] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c2aed5f3-d518-49d3-9e2d-630fd805cf85 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.252295] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-1f61d9fe-563d-45cf-80c0-eb3b6899f2fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1008.252609] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-1f61d9fe-563d-45cf-80c0-eb3b6899f2fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Deleting contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1008.252914] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f61d9fe-563d-45cf-80c0-eb3b6899f2fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Deleting the datastore file [datastore1] f23f2c04-c266-46a2-ad3f-34100f6246a0 {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1008.253305] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5faac0ba-fa2b-4add-aaa0-3ed0b36624ee {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.261446] env[63202]: DEBUG oslo_vmware.api [None req-1f61d9fe-563d-45cf-80c0-eb3b6899f2fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for the task: (returnval){ [ 1008.261446] env[63202]: value = "task-1385755" [ 1008.261446] env[63202]: _type = "Task" [ 1008.261446] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.272133] env[63202]: DEBUG oslo_vmware.api [None req-1f61d9fe-563d-45cf-80c0-eb3b6899f2fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385755, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.415334] env[63202]: DEBUG oslo_vmware.api [None req-937a425e-6dad-4384-9f38-0330ccbf0c18 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385753, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141689} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.416019] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-937a425e-6dad-4384-9f38-0330ccbf0c18 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1008.416019] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-937a425e-6dad-4384-9f38-0330ccbf0c18 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Deleted contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1008.416019] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-937a425e-6dad-4384-9f38-0330ccbf0c18 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1008.416190] env[63202]: INFO nova.compute.manager [None req-937a425e-6dad-4384-9f38-0330ccbf0c18 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Took 1.22 seconds to destroy the instance on the hypervisor. [ 1008.416362] env[63202]: DEBUG oslo.service.loopingcall [None req-937a425e-6dad-4384-9f38-0330ccbf0c18 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1008.416550] env[63202]: DEBUG nova.compute.manager [-] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1008.416636] env[63202]: DEBUG nova.network.neutron [-] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1008.493585] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1008.493845] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1008.494014] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Starting heal instance info cache {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1008.522455] env[63202]: DEBUG nova.compute.manager [req-ef3c4746-e4a3-4dc2-bfdd-4d7f1b9eb307 req-c5b874dd-0cd1-47b4-8424-7908a3fe78e1 service nova] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Received event network-changed-131d0949-eb5d-4ee6-a244-d7898f8e7b0f {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1008.522763] env[63202]: DEBUG nova.compute.manager [req-ef3c4746-e4a3-4dc2-bfdd-4d7f1b9eb307 req-c5b874dd-0cd1-47b4-8424-7908a3fe78e1 service nova] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Refreshing instance network info cache due to event network-changed-131d0949-eb5d-4ee6-a244-d7898f8e7b0f. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1008.522865] env[63202]: DEBUG oslo_concurrency.lockutils [req-ef3c4746-e4a3-4dc2-bfdd-4d7f1b9eb307 req-c5b874dd-0cd1-47b4-8424-7908a3fe78e1 service nova] Acquiring lock "refresh_cache-3c6d286d-c152-46f1-b212-96a67324a56d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.523029] env[63202]: DEBUG oslo_concurrency.lockutils [req-ef3c4746-e4a3-4dc2-bfdd-4d7f1b9eb307 req-c5b874dd-0cd1-47b4-8424-7908a3fe78e1 service nova] Acquired lock "refresh_cache-3c6d286d-c152-46f1-b212-96a67324a56d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.523194] env[63202]: DEBUG nova.network.neutron [req-ef3c4746-e4a3-4dc2-bfdd-4d7f1b9eb307 req-c5b874dd-0cd1-47b4-8424-7908a3fe78e1 service nova] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Refreshing network info cache for port 131d0949-eb5d-4ee6-a244-d7898f8e7b0f {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1008.553138] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c96bf224-4681-43e0-b8e2-89d1564d207f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "3c6d286d-c152-46f1-b212-96a67324a56d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.743s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.677903] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2233b51-bd88-4e4e-bd3b-8d747a28ddf3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.692053] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-016b2a02-d4c1-448e-b2e2-56fb6c449cbb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.696416] env[63202]: DEBUG nova.compute.manager [req-7c651813-5f51-4392-94f9-695cb22cbe14 req-52fa54e2-4ae8-4f82-8611-a0e2c49c9ff1 service nova] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Received event network-vif-deleted-e7f10c86-d09d-4e51-a8f7-deef174f2539 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1008.696625] env[63202]: INFO nova.compute.manager [req-7c651813-5f51-4392-94f9-695cb22cbe14 req-52fa54e2-4ae8-4f82-8611-a0e2c49c9ff1 service nova] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Neutron deleted interface e7f10c86-d09d-4e51-a8f7-deef174f2539; detaching it from the instance and deleting it from the info cache [ 1008.696792] env[63202]: DEBUG nova.network.neutron [req-7c651813-5f51-4392-94f9-695cb22cbe14 req-52fa54e2-4ae8-4f82-8611-a0e2c49c9ff1 service nova] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1008.730048] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d507fe7-9c49-4bd0-b4ee-e58acc7adab8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.738954] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34ffe14c-1495-4e26-a024-71d0898409ad {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.758097] env[63202]: DEBUG nova.compute.provider_tree [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Updating inventory in ProviderTree for provider 79b33d17-3e75-494c-a550-67b275de2079 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1008.771640] env[63202]: DEBUG oslo_vmware.api [None req-1f61d9fe-563d-45cf-80c0-eb3b6899f2fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Task: {'id': task-1385755, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.135539} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.774164] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f61d9fe-563d-45cf-80c0-eb3b6899f2fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1008.774365] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-1f61d9fe-563d-45cf-80c0-eb3b6899f2fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Deleted contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1008.774543] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-1f61d9fe-563d-45cf-80c0-eb3b6899f2fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1008.774714] env[63202]: INFO nova.compute.manager [None req-1f61d9fe-563d-45cf-80c0-eb3b6899f2fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1008.774954] env[63202]: DEBUG oslo.service.loopingcall [None req-1f61d9fe-563d-45cf-80c0-eb3b6899f2fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1008.775446] env[63202]: DEBUG nova.compute.manager [-] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1008.775540] env[63202]: DEBUG nova.network.neutron [-] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1008.813941] env[63202]: DEBUG oslo_concurrency.lockutils [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "dd146c71-c391-41e8-8cc7-7276f4a518d9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.814201] env[63202]: DEBUG oslo_concurrency.lockutils [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "dd146c71-c391-41e8-8cc7-7276f4a518d9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.814379] env[63202]: DEBUG oslo_concurrency.lockutils [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "dd146c71-c391-41e8-8cc7-7276f4a518d9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.973667] env[63202]: DEBUG nova.network.neutron [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Updating instance_info_cache with network_info: [{"id": "df7c236b-8be0-46d0-9dc1-56cba5b73ec7", "address": "fa:16:3e:93:8e:9b", "network": {"id": "5a3b4c9b-2ca7-4f8b-8bbb-ea000db91402", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-543677750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f508ec04edc844a19640a8a85f27e5b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf7c236b-8b", "ovs_interfaceid": "df7c236b-8be0-46d0-9dc1-56cba5b73ec7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.168052] env[63202]: DEBUG nova.network.neutron [-] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.199680] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6c9d1a0e-523c-4ddb-82ca-c2924f3b59c3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.212557] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1138875-84c9-439f-aaff-4c8ce024d048 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.241644] env[63202]: DEBUG nova.compute.manager [req-7c651813-5f51-4392-94f9-695cb22cbe14 req-52fa54e2-4ae8-4f82-8611-a0e2c49c9ff1 service nova] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Detach interface failed, port_id=e7f10c86-d09d-4e51-a8f7-deef174f2539, reason: Instance 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8 could not be found. {{(pid=63202) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1009.242906] env[63202]: DEBUG nova.network.neutron [req-ef3c4746-e4a3-4dc2-bfdd-4d7f1b9eb307 req-c5b874dd-0cd1-47b4-8424-7908a3fe78e1 service nova] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Updated VIF entry in instance network info cache for port 131d0949-eb5d-4ee6-a244-d7898f8e7b0f. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1009.243304] env[63202]: DEBUG nova.network.neutron [req-ef3c4746-e4a3-4dc2-bfdd-4d7f1b9eb307 req-c5b874dd-0cd1-47b4-8424-7908a3fe78e1 service nova] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Updating instance_info_cache with network_info: [{"id": "131d0949-eb5d-4ee6-a244-d7898f8e7b0f", "address": "fa:16:3e:c8:01:19", "network": {"id": "0aa55fee-953e-4c2d-b88a-43b0cf86c7f3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-878071414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.149", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7d28fea097541adadf0839940568409", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap131d0949-eb", "ovs_interfaceid": "131d0949-eb5d-4ee6-a244-d7898f8e7b0f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.292349] env[63202]: DEBUG nova.scheduler.client.report [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Updated inventory for provider 79b33d17-3e75-494c-a550-67b275de2079 with generation 112 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1009.292615] env[63202]: DEBUG nova.compute.provider_tree [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Updating resource provider 79b33d17-3e75-494c-a550-67b275de2079 generation from 112 to 113 during operation: update_inventory {{(pid=63202) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1009.292793] env[63202]: DEBUG nova.compute.provider_tree [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Updating inventory in ProviderTree for provider 79b33d17-3e75-494c-a550-67b275de2079 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1009.479113] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Releasing lock "refresh_cache-1bfb520e-6708-45a0-bb1f-2037fe1d801e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.499531] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Didn't find any instances for network info cache update. {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1009.499770] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1009.499934] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1009.500137] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1009.500281] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63202) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1009.513949] env[63202]: DEBUG nova.network.neutron [-] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.673375] env[63202]: INFO nova.compute.manager [-] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Took 1.26 seconds to deallocate network for instance. [ 1009.746043] env[63202]: DEBUG oslo_concurrency.lockutils [req-ef3c4746-e4a3-4dc2-bfdd-4d7f1b9eb307 req-c5b874dd-0cd1-47b4-8424-7908a3fe78e1 service nova] Releasing lock "refresh_cache-3c6d286d-c152-46f1-b212-96a67324a56d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.772029] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1009.772941] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef4404ca-3140-4bd0-bdca-3b85a7544a43 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.781536] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1009.781536] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8acceb2f-bae2-4181-9d87-08297db5460e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.797988] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.302s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.798515] env[63202]: DEBUG nova.compute.manager [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: b29a8043-2006-4255-8036-669d2be594d3] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1009.857642] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1009.857908] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Deleting contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1009.857908] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Deleting the datastore file [datastore1] 1bfb520e-6708-45a0-bb1f-2037fe1d801e {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1009.858161] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0702eece-6e57-425d-846f-b75e4227b18b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.866679] env[63202]: DEBUG oslo_vmware.api [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 1009.866679] env[63202]: value = "task-1385757" [ 1009.866679] env[63202]: _type = "Task" [ 1009.866679] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.871086] env[63202]: DEBUG oslo_concurrency.lockutils [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "refresh_cache-dd146c71-c391-41e8-8cc7-7276f4a518d9" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.871301] env[63202]: DEBUG oslo_concurrency.lockutils [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquired lock "refresh_cache-dd146c71-c391-41e8-8cc7-7276f4a518d9" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.871565] env[63202]: DEBUG nova.network.neutron [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1009.878045] env[63202]: DEBUG oslo_vmware.api [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385757, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.017354] env[63202]: INFO nova.compute.manager [-] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Took 1.24 seconds to deallocate network for instance. [ 1010.180452] env[63202]: DEBUG oslo_concurrency.lockutils [None req-937a425e-6dad-4384-9f38-0330ccbf0c18 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.180863] env[63202]: DEBUG oslo_concurrency.lockutils [None req-937a425e-6dad-4384-9f38-0330ccbf0c18 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.181135] env[63202]: DEBUG nova.objects.instance [None req-937a425e-6dad-4384-9f38-0330ccbf0c18 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Lazy-loading 'resources' on Instance uuid 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1010.254486] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1010.303785] env[63202]: DEBUG nova.compute.utils [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1010.305122] env[63202]: DEBUG nova.compute.manager [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: b29a8043-2006-4255-8036-669d2be594d3] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1010.305348] env[63202]: DEBUG nova.network.neutron [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: b29a8043-2006-4255-8036-669d2be594d3] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1010.342965] env[63202]: DEBUG nova.policy [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2bd91973e1bb4ce7be8bce9bfe34a941', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e972faeaedc6468aab7e7cfee88a477b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 1010.378808] env[63202]: DEBUG oslo_vmware.api [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385757, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.129231} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.379803] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1010.379803] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Deleted contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1010.379803] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1010.401716] env[63202]: INFO nova.scheduler.client.report [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Deleted allocations for instance 1bfb520e-6708-45a0-bb1f-2037fe1d801e [ 1010.527128] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1f61d9fe-563d-45cf-80c0-eb3b6899f2fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.595941] env[63202]: DEBUG nova.network.neutron [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Updating instance_info_cache with network_info: [{"id": "0bd2810b-f01e-4cd8-9c3e-76ac2339047a", "address": "fa:16:3e:9f:08:72", "network": {"id": "18390479-16ce-4012-9ba6-abf19b5616d1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1376026898-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22352c25bca8416a948014391a5389ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "085fb0ff-9285-4f1d-a008-a14da4844357", "external-id": "nsx-vlan-transportzone-729", "segmentation_id": 729, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0bd2810b-f0", "ovs_interfaceid": "0bd2810b-f01e-4cd8-9c3e-76ac2339047a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.610842] env[63202]: DEBUG nova.network.neutron [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: b29a8043-2006-4255-8036-669d2be594d3] Successfully created port: cefc0ade-cf96-4e99-8ebf-a64ae8190649 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1010.725509] env[63202]: DEBUG nova.compute.manager [req-c13d314b-407d-4acd-9efd-4d678701bcae req-363a94ce-a15f-4246-b447-b04a30521853 service nova] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Received event network-vif-deleted-e9c43ed4-c64a-47ae-9574-2985bd75f032 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1010.725790] env[63202]: DEBUG nova.compute.manager [req-c13d314b-407d-4acd-9efd-4d678701bcae req-363a94ce-a15f-4246-b447-b04a30521853 service nova] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Received event network-vif-unplugged-df7c236b-8be0-46d0-9dc1-56cba5b73ec7 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1010.726028] env[63202]: DEBUG oslo_concurrency.lockutils [req-c13d314b-407d-4acd-9efd-4d678701bcae req-363a94ce-a15f-4246-b447-b04a30521853 service nova] Acquiring lock "1bfb520e-6708-45a0-bb1f-2037fe1d801e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.726085] env[63202]: DEBUG oslo_concurrency.lockutils [req-c13d314b-407d-4acd-9efd-4d678701bcae req-363a94ce-a15f-4246-b447-b04a30521853 service nova] Lock "1bfb520e-6708-45a0-bb1f-2037fe1d801e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.726227] env[63202]: DEBUG oslo_concurrency.lockutils [req-c13d314b-407d-4acd-9efd-4d678701bcae req-363a94ce-a15f-4246-b447-b04a30521853 service nova] Lock "1bfb520e-6708-45a0-bb1f-2037fe1d801e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.726387] env[63202]: DEBUG nova.compute.manager [req-c13d314b-407d-4acd-9efd-4d678701bcae req-363a94ce-a15f-4246-b447-b04a30521853 service nova] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] No waiting events found dispatching network-vif-unplugged-df7c236b-8be0-46d0-9dc1-56cba5b73ec7 {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1010.726551] env[63202]: WARNING nova.compute.manager [req-c13d314b-407d-4acd-9efd-4d678701bcae req-363a94ce-a15f-4246-b447-b04a30521853 service nova] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Received unexpected event network-vif-unplugged-df7c236b-8be0-46d0-9dc1-56cba5b73ec7 for instance with vm_state shelved_offloaded and task_state None. [ 1010.726713] env[63202]: DEBUG nova.compute.manager [req-c13d314b-407d-4acd-9efd-4d678701bcae req-363a94ce-a15f-4246-b447-b04a30521853 service nova] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Received event network-changed-df7c236b-8be0-46d0-9dc1-56cba5b73ec7 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1010.726875] env[63202]: DEBUG nova.compute.manager [req-c13d314b-407d-4acd-9efd-4d678701bcae req-363a94ce-a15f-4246-b447-b04a30521853 service nova] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Refreshing instance network info cache due to event network-changed-df7c236b-8be0-46d0-9dc1-56cba5b73ec7. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1010.727075] env[63202]: DEBUG oslo_concurrency.lockutils [req-c13d314b-407d-4acd-9efd-4d678701bcae req-363a94ce-a15f-4246-b447-b04a30521853 service nova] Acquiring lock "refresh_cache-1bfb520e-6708-45a0-bb1f-2037fe1d801e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.727215] env[63202]: DEBUG oslo_concurrency.lockutils [req-c13d314b-407d-4acd-9efd-4d678701bcae req-363a94ce-a15f-4246-b447-b04a30521853 service nova] Acquired lock "refresh_cache-1bfb520e-6708-45a0-bb1f-2037fe1d801e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.727368] env[63202]: DEBUG nova.network.neutron [req-c13d314b-407d-4acd-9efd-4d678701bcae req-363a94ce-a15f-4246-b447-b04a30521853 service nova] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Refreshing network info cache for port df7c236b-8be0-46d0-9dc1-56cba5b73ec7 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1010.808097] env[63202]: DEBUG nova.compute.manager [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: b29a8043-2006-4255-8036-669d2be594d3] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1010.848776] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc596c0e-8896-42d8-b53f-94025c8339a3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.858366] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89408fce-fda2-4ada-8ac4-2174092d7876 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.891690] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b6c5743-6bd6-4e13-afba-0699666c3bab {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.900274] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ff77375-a4ba-499e-8ccd-1f782aacb7ea {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.905744] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.915082] env[63202]: DEBUG nova.compute.provider_tree [None req-937a425e-6dad-4384-9f38-0330ccbf0c18 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1011.099739] env[63202]: DEBUG oslo_concurrency.lockutils [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Releasing lock "refresh_cache-dd146c71-c391-41e8-8cc7-7276f4a518d9" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.418304] env[63202]: DEBUG nova.scheduler.client.report [None req-937a425e-6dad-4384-9f38-0330ccbf0c18 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1011.499032] env[63202]: DEBUG nova.network.neutron [req-c13d314b-407d-4acd-9efd-4d678701bcae req-363a94ce-a15f-4246-b447-b04a30521853 service nova] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Updated VIF entry in instance network info cache for port df7c236b-8be0-46d0-9dc1-56cba5b73ec7. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1011.499457] env[63202]: DEBUG nova.network.neutron [req-c13d314b-407d-4acd-9efd-4d678701bcae req-363a94ce-a15f-4246-b447-b04a30521853 service nova] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Updating instance_info_cache with network_info: [{"id": "df7c236b-8be0-46d0-9dc1-56cba5b73ec7", "address": "fa:16:3e:93:8e:9b", "network": {"id": "5a3b4c9b-2ca7-4f8b-8bbb-ea000db91402", "bridge": null, "label": "tempest-DeleteServersTestJSON-543677750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f508ec04edc844a19640a8a85f27e5b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapdf7c236b-8b", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.625449] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-764e2a6b-7c24-4d0d-9f83-57a89ffc8163 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.645881] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbacaf48-34cc-4ed7-b484-11816a26301e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.654170] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Updating instance 'dd146c71-c391-41e8-8cc7-7276f4a518d9' progress to 83 {{(pid=63202) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1011.774053] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89cdb25c-b0db-4081-b469-82e4b0e82c7f tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "1bfb520e-6708-45a0-bb1f-2037fe1d801e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.817027] env[63202]: DEBUG nova.compute.manager [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: b29a8043-2006-4255-8036-669d2be594d3] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1011.841146] env[63202]: DEBUG nova.virt.hardware [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1011.841407] env[63202]: DEBUG nova.virt.hardware [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1011.841563] env[63202]: DEBUG nova.virt.hardware [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1011.841765] env[63202]: DEBUG nova.virt.hardware [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1011.841901] env[63202]: DEBUG nova.virt.hardware [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1011.842056] env[63202]: DEBUG nova.virt.hardware [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1011.842269] env[63202]: DEBUG nova.virt.hardware [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1011.842832] env[63202]: DEBUG nova.virt.hardware [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1011.842832] env[63202]: DEBUG nova.virt.hardware [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1011.842832] env[63202]: DEBUG nova.virt.hardware [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1011.842999] env[63202]: DEBUG nova.virt.hardware [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1011.843761] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c7a51c3-6403-481d-8b64-b7de715eb6a7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.853840] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84f17fc2-4822-492b-888b-376b560541af {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.924986] env[63202]: DEBUG oslo_concurrency.lockutils [None req-937a425e-6dad-4384-9f38-0330ccbf0c18 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.744s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.928163] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1f61d9fe-563d-45cf-80c0-eb3b6899f2fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.400s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.928393] env[63202]: DEBUG nova.objects.instance [None req-1f61d9fe-563d-45cf-80c0-eb3b6899f2fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lazy-loading 'resources' on Instance uuid f23f2c04-c266-46a2-ad3f-34100f6246a0 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1011.942659] env[63202]: INFO nova.scheduler.client.report [None req-937a425e-6dad-4384-9f38-0330ccbf0c18 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Deleted allocations for instance 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8 [ 1012.001800] env[63202]: DEBUG oslo_concurrency.lockutils [req-c13d314b-407d-4acd-9efd-4d678701bcae req-363a94ce-a15f-4246-b447-b04a30521853 service nova] Releasing lock "refresh_cache-1bfb520e-6708-45a0-bb1f-2037fe1d801e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.102753] env[63202]: DEBUG nova.network.neutron [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: b29a8043-2006-4255-8036-669d2be594d3] Successfully updated port: cefc0ade-cf96-4e99-8ebf-a64ae8190649 {{(pid=63202) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1012.160239] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1012.160548] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-777277d8-cfcf-4758-b6d7-9a3873a05d3c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.168398] env[63202]: DEBUG oslo_vmware.api [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for the task: (returnval){ [ 1012.168398] env[63202]: value = "task-1385758" [ 1012.168398] env[63202]: _type = "Task" [ 1012.168398] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.177045] env[63202]: DEBUG oslo_vmware.api [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385758, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.451029] env[63202]: DEBUG oslo_concurrency.lockutils [None req-937a425e-6dad-4384-9f38-0330ccbf0c18 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Lock "8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.260s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.565688] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-273aa025-1709-4bea-8bbc-64b0b43e9cc5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.573936] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96453a53-3498-4a7b-8d10-5059c2dd7b4c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.603918] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-902274c5-6441-4785-9f13-4f59d8f9ac51 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.606551] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "refresh_cache-b29a8043-2006-4255-8036-669d2be594d3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.606713] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquired lock "refresh_cache-b29a8043-2006-4255-8036-669d2be594d3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.606886] env[63202]: DEBUG nova.network.neutron [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: b29a8043-2006-4255-8036-669d2be594d3] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1012.613037] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a79079f0-e559-4f7d-8d84-3b4b17e7a06a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.629171] env[63202]: DEBUG nova.compute.provider_tree [None req-1f61d9fe-563d-45cf-80c0-eb3b6899f2fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1012.679149] env[63202]: DEBUG oslo_vmware.api [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385758, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.751075] env[63202]: DEBUG nova.compute.manager [req-e7276250-12a7-4569-97e8-a2a6a27da67f req-1450720e-411a-4898-a66b-02314e5f9923 service nova] [instance: b29a8043-2006-4255-8036-669d2be594d3] Received event network-vif-plugged-cefc0ade-cf96-4e99-8ebf-a64ae8190649 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1012.751318] env[63202]: DEBUG oslo_concurrency.lockutils [req-e7276250-12a7-4569-97e8-a2a6a27da67f req-1450720e-411a-4898-a66b-02314e5f9923 service nova] Acquiring lock "b29a8043-2006-4255-8036-669d2be594d3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1012.751615] env[63202]: DEBUG oslo_concurrency.lockutils [req-e7276250-12a7-4569-97e8-a2a6a27da67f req-1450720e-411a-4898-a66b-02314e5f9923 service nova] Lock "b29a8043-2006-4255-8036-669d2be594d3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1012.751812] env[63202]: DEBUG oslo_concurrency.lockutils [req-e7276250-12a7-4569-97e8-a2a6a27da67f req-1450720e-411a-4898-a66b-02314e5f9923 service nova] Lock "b29a8043-2006-4255-8036-669d2be594d3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.751993] env[63202]: DEBUG nova.compute.manager [req-e7276250-12a7-4569-97e8-a2a6a27da67f req-1450720e-411a-4898-a66b-02314e5f9923 service nova] [instance: b29a8043-2006-4255-8036-669d2be594d3] No waiting events found dispatching network-vif-plugged-cefc0ade-cf96-4e99-8ebf-a64ae8190649 {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1012.752249] env[63202]: WARNING nova.compute.manager [req-e7276250-12a7-4569-97e8-a2a6a27da67f req-1450720e-411a-4898-a66b-02314e5f9923 service nova] [instance: b29a8043-2006-4255-8036-669d2be594d3] Received unexpected event network-vif-plugged-cefc0ade-cf96-4e99-8ebf-a64ae8190649 for instance with vm_state building and task_state spawning. [ 1012.752353] env[63202]: DEBUG nova.compute.manager [req-e7276250-12a7-4569-97e8-a2a6a27da67f req-1450720e-411a-4898-a66b-02314e5f9923 service nova] [instance: b29a8043-2006-4255-8036-669d2be594d3] Received event network-changed-cefc0ade-cf96-4e99-8ebf-a64ae8190649 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1012.752515] env[63202]: DEBUG nova.compute.manager [req-e7276250-12a7-4569-97e8-a2a6a27da67f req-1450720e-411a-4898-a66b-02314e5f9923 service nova] [instance: b29a8043-2006-4255-8036-669d2be594d3] Refreshing instance network info cache due to event network-changed-cefc0ade-cf96-4e99-8ebf-a64ae8190649. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1012.753355] env[63202]: DEBUG oslo_concurrency.lockutils [req-e7276250-12a7-4569-97e8-a2a6a27da67f req-1450720e-411a-4898-a66b-02314e5f9923 service nova] Acquiring lock "refresh_cache-b29a8043-2006-4255-8036-669d2be594d3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.108036] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Acquiring lock "3029c06f-4687-46f7-958b-831e4d0b241b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.108359] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Lock "3029c06f-4687-46f7-958b-831e4d0b241b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.132244] env[63202]: DEBUG nova.scheduler.client.report [None req-1f61d9fe-563d-45cf-80c0-eb3b6899f2fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1013.138953] env[63202]: DEBUG nova.network.neutron [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: b29a8043-2006-4255-8036-669d2be594d3] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1013.178471] env[63202]: DEBUG oslo_vmware.api [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385758, 'name': PowerOnVM_Task, 'duration_secs': 0.568991} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.180721] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1013.180942] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-64d442a6-5eb7-4137-b9dd-713bffc59428 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Updating instance 'dd146c71-c391-41e8-8cc7-7276f4a518d9' progress to 100 {{(pid=63202) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1013.282121] env[63202]: DEBUG nova.network.neutron [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: b29a8043-2006-4255-8036-669d2be594d3] Updating instance_info_cache with network_info: [{"id": "cefc0ade-cf96-4e99-8ebf-a64ae8190649", "address": "fa:16:3e:fd:32:a5", "network": {"id": "b1ea2320-3801-43fe-b566-4f69db81329e", "bridge": "br-int", "label": "tempest-ServersTestJSON-618582355-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e972faeaedc6468aab7e7cfee88a477b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4223acd2-30f7-440e-b975-60b30d931694", "external-id": "nsx-vlan-transportzone-647", "segmentation_id": 647, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcefc0ade-cf", "ovs_interfaceid": "cefc0ade-cf96-4e99-8ebf-a64ae8190649", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1013.612615] env[63202]: DEBUG nova.compute.manager [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1013.638373] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1f61d9fe-563d-45cf-80c0-eb3b6899f2fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.711s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.640458] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.735s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.640695] env[63202]: DEBUG nova.objects.instance [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lazy-loading 'resources' on Instance uuid 1bfb520e-6708-45a0-bb1f-2037fe1d801e {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1013.660108] env[63202]: INFO nova.scheduler.client.report [None req-1f61d9fe-563d-45cf-80c0-eb3b6899f2fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Deleted allocations for instance f23f2c04-c266-46a2-ad3f-34100f6246a0 [ 1013.785075] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Releasing lock "refresh_cache-b29a8043-2006-4255-8036-669d2be594d3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.785373] env[63202]: DEBUG nova.compute.manager [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: b29a8043-2006-4255-8036-669d2be594d3] Instance network_info: |[{"id": "cefc0ade-cf96-4e99-8ebf-a64ae8190649", "address": "fa:16:3e:fd:32:a5", "network": {"id": "b1ea2320-3801-43fe-b566-4f69db81329e", "bridge": "br-int", "label": "tempest-ServersTestJSON-618582355-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e972faeaedc6468aab7e7cfee88a477b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4223acd2-30f7-440e-b975-60b30d931694", "external-id": "nsx-vlan-transportzone-647", "segmentation_id": 647, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcefc0ade-cf", "ovs_interfaceid": "cefc0ade-cf96-4e99-8ebf-a64ae8190649", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1013.785687] env[63202]: DEBUG oslo_concurrency.lockutils [req-e7276250-12a7-4569-97e8-a2a6a27da67f req-1450720e-411a-4898-a66b-02314e5f9923 service nova] Acquired lock "refresh_cache-b29a8043-2006-4255-8036-669d2be594d3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.785924] env[63202]: DEBUG nova.network.neutron [req-e7276250-12a7-4569-97e8-a2a6a27da67f req-1450720e-411a-4898-a66b-02314e5f9923 service nova] [instance: b29a8043-2006-4255-8036-669d2be594d3] Refreshing network info cache for port cefc0ade-cf96-4e99-8ebf-a64ae8190649 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1013.787500] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: b29a8043-2006-4255-8036-669d2be594d3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fd:32:a5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4223acd2-30f7-440e-b975-60b30d931694', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cefc0ade-cf96-4e99-8ebf-a64ae8190649', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1013.796468] env[63202]: DEBUG oslo.service.loopingcall [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1013.797499] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b29a8043-2006-4255-8036-669d2be594d3] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1013.797739] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-61defd1e-41cf-49b4-86c2-b59226186a18 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.819763] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1013.819763] env[63202]: value = "task-1385759" [ 1013.819763] env[63202]: _type = "Task" [ 1013.819763] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.830404] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385759, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.135355] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.143729] env[63202]: DEBUG nova.objects.instance [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lazy-loading 'numa_topology' on Instance uuid 1bfb520e-6708-45a0-bb1f-2037fe1d801e {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1014.167670] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1f61d9fe-563d-45cf-80c0-eb3b6899f2fa tempest-ImagesTestJSON-120260003 tempest-ImagesTestJSON-120260003-project-member] Lock "f23f2c04-c266-46a2-ad3f-34100f6246a0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.545s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.330723] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385759, 'name': CreateVM_Task, 'duration_secs': 0.282356} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.330926] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b29a8043-2006-4255-8036-669d2be594d3] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1014.331631] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.331832] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.332210] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1014.332478] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8101ff8e-1cd9-4a62-a02a-6d6a04916256 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.337970] env[63202]: DEBUG oslo_vmware.api [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 1014.337970] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]521dfb9d-add3-17c6-7a9b-e5abd0ab27e1" [ 1014.337970] env[63202]: _type = "Task" [ 1014.337970] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.346961] env[63202]: DEBUG oslo_vmware.api [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]521dfb9d-add3-17c6-7a9b-e5abd0ab27e1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.647237] env[63202]: DEBUG nova.objects.base [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Object Instance<1bfb520e-6708-45a0-bb1f-2037fe1d801e> lazy-loaded attributes: resources,numa_topology {{(pid=63202) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1014.659393] env[63202]: DEBUG nova.network.neutron [req-e7276250-12a7-4569-97e8-a2a6a27da67f req-1450720e-411a-4898-a66b-02314e5f9923 service nova] [instance: b29a8043-2006-4255-8036-669d2be594d3] Updated VIF entry in instance network info cache for port cefc0ade-cf96-4e99-8ebf-a64ae8190649. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1014.659726] env[63202]: DEBUG nova.network.neutron [req-e7276250-12a7-4569-97e8-a2a6a27da67f req-1450720e-411a-4898-a66b-02314e5f9923 service nova] [instance: b29a8043-2006-4255-8036-669d2be594d3] Updating instance_info_cache with network_info: [{"id": "cefc0ade-cf96-4e99-8ebf-a64ae8190649", "address": "fa:16:3e:fd:32:a5", "network": {"id": "b1ea2320-3801-43fe-b566-4f69db81329e", "bridge": "br-int", "label": "tempest-ServersTestJSON-618582355-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e972faeaedc6468aab7e7cfee88a477b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4223acd2-30f7-440e-b975-60b30d931694", "external-id": "nsx-vlan-transportzone-647", "segmentation_id": 647, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcefc0ade-cf", "ovs_interfaceid": "cefc0ade-cf96-4e99-8ebf-a64ae8190649", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.799788] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d165e421-cc92-4a48-b036-4f97b3e8af26 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.809851] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cfa61a7-1f82-4d2d-b511-e9f328317255 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.849108] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a1debd3-e3c1-4a75-99f4-4b3e6e3fef31 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.859050] env[63202]: DEBUG oslo_vmware.api [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]521dfb9d-add3-17c6-7a9b-e5abd0ab27e1, 'name': SearchDatastore_Task, 'duration_secs': 0.010401} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.861824] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.862104] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: b29a8043-2006-4255-8036-669d2be594d3] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1014.862364] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.862514] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.862698] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1014.863074] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f9539ac0-2634-4157-8ab1-fd58514ecc37 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.866213] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0474c2dc-3290-43f0-a69b-bf7fcc92ea9a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.882500] env[63202]: DEBUG nova.compute.provider_tree [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1014.885314] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1014.885555] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1014.886520] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2b35c90e-6218-4923-af3d-75690592ab34 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.893583] env[63202]: DEBUG oslo_vmware.api [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 1014.893583] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52bd7da0-9f85-4038-ac80-ace6d9b79d73" [ 1014.893583] env[63202]: _type = "Task" [ 1014.893583] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.903922] env[63202]: DEBUG oslo_vmware.api [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52bd7da0-9f85-4038-ac80-ace6d9b79d73, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.163287] env[63202]: DEBUG oslo_concurrency.lockutils [req-e7276250-12a7-4569-97e8-a2a6a27da67f req-1450720e-411a-4898-a66b-02314e5f9923 service nova] Releasing lock "refresh_cache-b29a8043-2006-4255-8036-669d2be594d3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.180353] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3563b9ec-752c-4f6d-b5c5-e46a3a94bf81 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "dd146c71-c391-41e8-8cc7-7276f4a518d9" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.180606] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3563b9ec-752c-4f6d-b5c5-e46a3a94bf81 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "dd146c71-c391-41e8-8cc7-7276f4a518d9" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.180967] env[63202]: DEBUG nova.compute.manager [None req-3563b9ec-752c-4f6d-b5c5-e46a3a94bf81 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Going to confirm migration 2 {{(pid=63202) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1015.387796] env[63202]: DEBUG nova.scheduler.client.report [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1015.404786] env[63202]: DEBUG oslo_vmware.api [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52bd7da0-9f85-4038-ac80-ace6d9b79d73, 'name': SearchDatastore_Task, 'duration_secs': 0.011558} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.406295] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ba0a9d5e-7168-44ca-b9a9-ff6f4e7a23ba {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.413334] env[63202]: DEBUG oslo_vmware.api [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 1015.413334] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]522adf02-5303-faea-01e1-a3a781ab94cd" [ 1015.413334] env[63202]: _type = "Task" [ 1015.413334] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.422939] env[63202]: DEBUG oslo_vmware.api [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]522adf02-5303-faea-01e1-a3a781ab94cd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.765400] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3563b9ec-752c-4f6d-b5c5-e46a3a94bf81 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "refresh_cache-dd146c71-c391-41e8-8cc7-7276f4a518d9" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.765724] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3563b9ec-752c-4f6d-b5c5-e46a3a94bf81 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquired lock "refresh_cache-dd146c71-c391-41e8-8cc7-7276f4a518d9" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.766017] env[63202]: DEBUG nova.network.neutron [None req-3563b9ec-752c-4f6d-b5c5-e46a3a94bf81 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1015.766344] env[63202]: DEBUG nova.objects.instance [None req-3563b9ec-752c-4f6d-b5c5-e46a3a94bf81 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lazy-loading 'info_cache' on Instance uuid dd146c71-c391-41e8-8cc7-7276f4a518d9 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1015.892883] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.252s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.895590] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.760s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.897098] env[63202]: INFO nova.compute.claims [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1015.925842] env[63202]: DEBUG oslo_vmware.api [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]522adf02-5303-faea-01e1-a3a781ab94cd, 'name': SearchDatastore_Task, 'duration_secs': 0.015686} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.925842] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.926077] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] b29a8043-2006-4255-8036-669d2be594d3/b29a8043-2006-4255-8036-669d2be594d3.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1015.926241] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2c0099f7-eabd-4e7c-9497-35cd06fb8a04 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.935124] env[63202]: DEBUG oslo_vmware.api [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 1015.935124] env[63202]: value = "task-1385760" [ 1015.935124] env[63202]: _type = "Task" [ 1015.935124] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.944625] env[63202]: DEBUG oslo_vmware.api [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385760, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.408928] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1ef6804f-fd15-42ed-be71-3db7e413fecb tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "1bfb520e-6708-45a0-bb1f-2037fe1d801e" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 26.309s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.410026] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89cdb25c-b0db-4081-b469-82e4b0e82c7f tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "1bfb520e-6708-45a0-bb1f-2037fe1d801e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 4.636s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.410185] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89cdb25c-b0db-4081-b469-82e4b0e82c7f tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "1bfb520e-6708-45a0-bb1f-2037fe1d801e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.410405] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89cdb25c-b0db-4081-b469-82e4b0e82c7f tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "1bfb520e-6708-45a0-bb1f-2037fe1d801e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.410583] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89cdb25c-b0db-4081-b469-82e4b0e82c7f tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "1bfb520e-6708-45a0-bb1f-2037fe1d801e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.421292] env[63202]: INFO nova.compute.manager [None req-89cdb25c-b0db-4081-b469-82e4b0e82c7f tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Terminating instance [ 1016.423347] env[63202]: DEBUG nova.compute.manager [None req-89cdb25c-b0db-4081-b469-82e4b0e82c7f tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1016.423658] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-89cdb25c-b0db-4081-b469-82e4b0e82c7f tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1016.423875] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-93751f01-9a90-4477-a047-81fe37b9fa9b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.435724] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6644aec7-8acb-49c1-87b1-7ebdbb80db4b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.459630] env[63202]: DEBUG oslo_vmware.api [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385760, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.472068] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-89cdb25c-b0db-4081-b469-82e4b0e82c7f tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1bfb520e-6708-45a0-bb1f-2037fe1d801e could not be found. [ 1016.472369] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-89cdb25c-b0db-4081-b469-82e4b0e82c7f tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1016.472571] env[63202]: INFO nova.compute.manager [None req-89cdb25c-b0db-4081-b469-82e4b0e82c7f tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Took 0.05 seconds to destroy the instance on the hypervisor. [ 1016.472857] env[63202]: DEBUG oslo.service.loopingcall [None req-89cdb25c-b0db-4081-b469-82e4b0e82c7f tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1016.473330] env[63202]: DEBUG nova.compute.manager [-] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1016.473444] env[63202]: DEBUG nova.network.neutron [-] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1016.961364] env[63202]: DEBUG oslo_vmware.api [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385760, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.88863} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.961505] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] b29a8043-2006-4255-8036-669d2be594d3/b29a8043-2006-4255-8036-669d2be594d3.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1016.961725] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: b29a8043-2006-4255-8036-669d2be594d3] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1016.962386] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bc4bf165-4f21-480a-b43c-a80218e2cc0b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.970475] env[63202]: DEBUG oslo_vmware.api [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 1016.970475] env[63202]: value = "task-1385761" [ 1016.970475] env[63202]: _type = "Task" [ 1016.970475] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.979206] env[63202]: DEBUG oslo_vmware.api [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385761, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.997927] env[63202]: DEBUG nova.network.neutron [None req-3563b9ec-752c-4f6d-b5c5-e46a3a94bf81 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Updating instance_info_cache with network_info: [{"id": "0bd2810b-f01e-4cd8-9c3e-76ac2339047a", "address": "fa:16:3e:9f:08:72", "network": {"id": "18390479-16ce-4012-9ba6-abf19b5616d1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1376026898-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22352c25bca8416a948014391a5389ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "085fb0ff-9285-4f1d-a008-a14da4844357", "external-id": "nsx-vlan-transportzone-729", "segmentation_id": 729, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0bd2810b-f0", "ovs_interfaceid": "0bd2810b-f01e-4cd8-9c3e-76ac2339047a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.041089] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-421e463b-85e5-45c1-8197-c8b8eac0bce0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.047815] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21d6cf19-647d-4ef8-8142-ccd2e3b4a9bd {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.081767] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-991feda1-0289-484c-8a38-36d4686ec31b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.090717] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c04ef4ae-a2ac-4b64-b4be-81e5c3058ba0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.104793] env[63202]: DEBUG nova.compute.provider_tree [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1017.248167] env[63202]: DEBUG nova.network.neutron [-] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.483723] env[63202]: DEBUG oslo_vmware.api [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385761, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076499} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.484088] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: b29a8043-2006-4255-8036-669d2be594d3] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1017.484815] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38e38eb9-23b2-4c52-8d69-be646e15605b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.507058] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: b29a8043-2006-4255-8036-669d2be594d3] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] b29a8043-2006-4255-8036-669d2be594d3/b29a8043-2006-4255-8036-669d2be594d3.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1017.507544] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3563b9ec-752c-4f6d-b5c5-e46a3a94bf81 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Releasing lock "refresh_cache-dd146c71-c391-41e8-8cc7-7276f4a518d9" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.507769] env[63202]: DEBUG nova.objects.instance [None req-3563b9ec-752c-4f6d-b5c5-e46a3a94bf81 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lazy-loading 'migration_context' on Instance uuid dd146c71-c391-41e8-8cc7-7276f4a518d9 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1017.508812] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b714cd50-ce3c-4b3d-b238-acf919edaf9e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.529937] env[63202]: DEBUG oslo_vmware.api [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 1017.529937] env[63202]: value = "task-1385762" [ 1017.529937] env[63202]: _type = "Task" [ 1017.529937] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.542474] env[63202]: DEBUG oslo_vmware.api [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385762, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.608685] env[63202]: DEBUG nova.scheduler.client.report [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1017.751170] env[63202]: INFO nova.compute.manager [-] [instance: 1bfb520e-6708-45a0-bb1f-2037fe1d801e] Took 1.28 seconds to deallocate network for instance. [ 1018.024647] env[63202]: DEBUG nova.objects.base [None req-3563b9ec-752c-4f6d-b5c5-e46a3a94bf81 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=63202) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1018.025581] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8473de1-8369-483e-a4c6-682d1a6265b4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.051142] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-151abc92-b565-4d04-a0a5-f8239e831838 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.056601] env[63202]: DEBUG oslo_vmware.api [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385762, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.059820] env[63202]: DEBUG oslo_vmware.api [None req-3563b9ec-752c-4f6d-b5c5-e46a3a94bf81 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for the task: (returnval){ [ 1018.059820] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]520ef582-9999-d538-a76e-370f6ddf64f9" [ 1018.059820] env[63202]: _type = "Task" [ 1018.059820] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.068833] env[63202]: DEBUG oslo_vmware.api [None req-3563b9ec-752c-4f6d-b5c5-e46a3a94bf81 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]520ef582-9999-d538-a76e-370f6ddf64f9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.115022] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.218s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.115022] env[63202]: DEBUG nova.compute.manager [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1018.541821] env[63202]: DEBUG oslo_vmware.api [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385762, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.568812] env[63202]: DEBUG oslo_vmware.api [None req-3563b9ec-752c-4f6d-b5c5-e46a3a94bf81 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]520ef582-9999-d538-a76e-370f6ddf64f9, 'name': SearchDatastore_Task, 'duration_secs': 0.030731} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.569096] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3563b9ec-752c-4f6d-b5c5-e46a3a94bf81 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.569360] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3563b9ec-752c-4f6d-b5c5-e46a3a94bf81 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.620019] env[63202]: DEBUG nova.compute.utils [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1018.621383] env[63202]: DEBUG nova.compute.manager [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1018.621543] env[63202]: DEBUG nova.network.neutron [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1018.659051] env[63202]: DEBUG nova.policy [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2b9d1001300c476a896f335b3acf041b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f651d18a9559499da27f301ee75bb2fb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 1018.779358] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89cdb25c-b0db-4081-b469-82e4b0e82c7f tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "1bfb520e-6708-45a0-bb1f-2037fe1d801e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.369s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.793059] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquiring lock "8437f856-a707-49c5-b8eb-5a22cdb990f8" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.793295] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lock "8437f856-a707-49c5-b8eb-5a22cdb990f8" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.793472] env[63202]: INFO nova.compute.manager [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Shelving [ 1018.911050] env[63202]: DEBUG nova.network.neutron [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Successfully created port: 44f7adb9-2f63-4a9a-a041-0a0ae57c78c8 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1019.042016] env[63202]: DEBUG oslo_vmware.api [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385762, 'name': ReconfigVM_Task, 'duration_secs': 1.202882} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.042323] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: b29a8043-2006-4255-8036-669d2be594d3] Reconfigured VM instance instance-0000005d to attach disk [datastore1] b29a8043-2006-4255-8036-669d2be594d3/b29a8043-2006-4255-8036-669d2be594d3.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1019.042944] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6a9d9a84-c527-4500-be91-053978b6a6d2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.050680] env[63202]: DEBUG oslo_vmware.api [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 1019.050680] env[63202]: value = "task-1385763" [ 1019.050680] env[63202]: _type = "Task" [ 1019.050680] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.058857] env[63202]: DEBUG oslo_vmware.api [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385763, 'name': Rename_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.125857] env[63202]: DEBUG nova.compute.manager [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1019.203476] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8341c658-392e-4694-b92b-8e4bc737b691 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.212071] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32645c79-ef83-425a-a8be-bffa77dda279 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.247918] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-feab6908-7248-40a4-bb26-7b0fb762b77e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.256013] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0b249b8-0dbf-47b3-bdbf-9ca57982cc8d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.270074] env[63202]: DEBUG nova.compute.provider_tree [None req-3563b9ec-752c-4f6d-b5c5-e46a3a94bf81 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1019.300283] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1019.300498] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-070d6311-6ad1-493b-b901-61981fe15a14 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.308894] env[63202]: DEBUG oslo_vmware.api [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1019.308894] env[63202]: value = "task-1385764" [ 1019.308894] env[63202]: _type = "Task" [ 1019.308894] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.317673] env[63202]: DEBUG oslo_vmware.api [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385764, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.560949] env[63202]: DEBUG oslo_vmware.api [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385763, 'name': Rename_Task, 'duration_secs': 0.191441} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.561396] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: b29a8043-2006-4255-8036-669d2be594d3] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1019.561396] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d2e80f93-dab7-4654-8375-516c8d59fc14 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.570813] env[63202]: DEBUG oslo_vmware.api [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 1019.570813] env[63202]: value = "task-1385765" [ 1019.570813] env[63202]: _type = "Task" [ 1019.570813] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.578425] env[63202]: DEBUG oslo_vmware.api [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385765, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.773478] env[63202]: DEBUG nova.scheduler.client.report [None req-3563b9ec-752c-4f6d-b5c5-e46a3a94bf81 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1019.819038] env[63202]: DEBUG oslo_vmware.api [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385764, 'name': PowerOffVM_Task, 'duration_secs': 0.222037} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.819332] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1019.820145] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaaabd7a-8e0e-4f4b-a094-9a437a514437 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.840724] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd1ab8d1-1436-4af7-975c-be1fe6128a1c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.081376] env[63202]: DEBUG oslo_vmware.api [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385765, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.134782] env[63202]: DEBUG nova.compute.manager [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1020.159190] env[63202]: DEBUG nova.virt.hardware [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1020.159458] env[63202]: DEBUG nova.virt.hardware [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1020.159618] env[63202]: DEBUG nova.virt.hardware [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1020.159799] env[63202]: DEBUG nova.virt.hardware [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1020.159947] env[63202]: DEBUG nova.virt.hardware [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1020.160118] env[63202]: DEBUG nova.virt.hardware [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1020.160331] env[63202]: DEBUG nova.virt.hardware [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1020.160490] env[63202]: DEBUG nova.virt.hardware [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1020.160655] env[63202]: DEBUG nova.virt.hardware [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1020.160838] env[63202]: DEBUG nova.virt.hardware [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1020.161044] env[63202]: DEBUG nova.virt.hardware [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1020.161976] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d5b0f03-084d-4e98-b16c-7a7cd3d10fff {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.169021] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.169281] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.174711] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e20b39fd-e410-4051-b68e-3435a56b9656 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.294183] env[63202]: DEBUG nova.compute.manager [req-8eaa907f-5800-46b2-9144-8d1346b959e4 req-df99649b-a82c-4cd7-876a-52e6488e6f39 service nova] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Received event network-vif-plugged-44f7adb9-2f63-4a9a-a041-0a0ae57c78c8 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1020.294183] env[63202]: DEBUG oslo_concurrency.lockutils [req-8eaa907f-5800-46b2-9144-8d1346b959e4 req-df99649b-a82c-4cd7-876a-52e6488e6f39 service nova] Acquiring lock "3029c06f-4687-46f7-958b-831e4d0b241b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.294183] env[63202]: DEBUG oslo_concurrency.lockutils [req-8eaa907f-5800-46b2-9144-8d1346b959e4 req-df99649b-a82c-4cd7-876a-52e6488e6f39 service nova] Lock "3029c06f-4687-46f7-958b-831e4d0b241b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.294183] env[63202]: DEBUG oslo_concurrency.lockutils [req-8eaa907f-5800-46b2-9144-8d1346b959e4 req-df99649b-a82c-4cd7-876a-52e6488e6f39 service nova] Lock "3029c06f-4687-46f7-958b-831e4d0b241b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.294183] env[63202]: DEBUG nova.compute.manager [req-8eaa907f-5800-46b2-9144-8d1346b959e4 req-df99649b-a82c-4cd7-876a-52e6488e6f39 service nova] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] No waiting events found dispatching network-vif-plugged-44f7adb9-2f63-4a9a-a041-0a0ae57c78c8 {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1020.294452] env[63202]: WARNING nova.compute.manager [req-8eaa907f-5800-46b2-9144-8d1346b959e4 req-df99649b-a82c-4cd7-876a-52e6488e6f39 service nova] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Received unexpected event network-vif-plugged-44f7adb9-2f63-4a9a-a041-0a0ae57c78c8 for instance with vm_state building and task_state spawning. [ 1020.352984] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Creating Snapshot of the VM instance {{(pid=63202) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1020.353326] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-055730d4-a817-4558-ba3c-0cb449391839 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.363864] env[63202]: DEBUG oslo_vmware.api [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1020.363864] env[63202]: value = "task-1385766" [ 1020.363864] env[63202]: _type = "Task" [ 1020.363864] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.375542] env[63202]: DEBUG oslo_vmware.api [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385766, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.398161] env[63202]: DEBUG nova.network.neutron [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Successfully updated port: 44f7adb9-2f63-4a9a-a041-0a0ae57c78c8 {{(pid=63202) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1020.581901] env[63202]: DEBUG oslo_vmware.api [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385765, 'name': PowerOnVM_Task, 'duration_secs': 0.586699} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.582243] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: b29a8043-2006-4255-8036-669d2be594d3] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1020.582372] env[63202]: INFO nova.compute.manager [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: b29a8043-2006-4255-8036-669d2be594d3] Took 8.77 seconds to spawn the instance on the hypervisor. [ 1020.582556] env[63202]: DEBUG nova.compute.manager [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: b29a8043-2006-4255-8036-669d2be594d3] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1020.583379] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61e04829-7ead-4d50-b58c-8f6c28915316 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.671689] env[63202]: DEBUG nova.compute.manager [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1020.784641] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3563b9ec-752c-4f6d-b5c5-e46a3a94bf81 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.215s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.874654] env[63202]: DEBUG oslo_vmware.api [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385766, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.900840] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Acquiring lock "refresh_cache-3029c06f-4687-46f7-958b-831e4d0b241b" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.901058] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Acquired lock "refresh_cache-3029c06f-4687-46f7-958b-831e4d0b241b" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.901190] env[63202]: DEBUG nova.network.neutron [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1021.099832] env[63202]: INFO nova.compute.manager [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: b29a8043-2006-4255-8036-669d2be594d3] Took 17.25 seconds to build instance. [ 1021.194999] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.195445] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.197771] env[63202]: INFO nova.compute.claims [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1021.339013] env[63202]: INFO nova.scheduler.client.report [None req-3563b9ec-752c-4f6d-b5c5-e46a3a94bf81 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Deleted allocation for migration 839330ab-3902-4a6d-965a-18d61af14cb7 [ 1021.376024] env[63202]: DEBUG oslo_vmware.api [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385766, 'name': CreateSnapshot_Task, 'duration_secs': 0.995621} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.376881] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Created Snapshot of the VM instance {{(pid=63202) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1021.377680] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9339834f-37c7-4eec-a0b1-869d113e369f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.432618] env[63202]: DEBUG nova.network.neutron [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1021.569805] env[63202]: DEBUG nova.network.neutron [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Updating instance_info_cache with network_info: [{"id": "44f7adb9-2f63-4a9a-a041-0a0ae57c78c8", "address": "fa:16:3e:88:33:11", "network": {"id": "69295efd-3c50-45e5-bc09-4f0cef6b8f0e", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1062402518-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f651d18a9559499da27f301ee75bb2fb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dced2f3d-7fd3-4a42-836d-9f02dab4c949", "external-id": "nsx-vlan-transportzone-117", "segmentation_id": 117, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44f7adb9-2f", "ovs_interfaceid": "44f7adb9-2f63-4a9a-a041-0a0ae57c78c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.602429] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a2fca271-5a44-4a41-9c36-4f19ebdf1548 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "b29a8043-2006-4255-8036-669d2be594d3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.765s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.814574] env[63202]: DEBUG oslo_concurrency.lockutils [None req-610d9e96-076b-4d48-a965-36d8cf6a8fa8 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "b29a8043-2006-4255-8036-669d2be594d3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.814835] env[63202]: DEBUG oslo_concurrency.lockutils [None req-610d9e96-076b-4d48-a965-36d8cf6a8fa8 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "b29a8043-2006-4255-8036-669d2be594d3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.815063] env[63202]: DEBUG oslo_concurrency.lockutils [None req-610d9e96-076b-4d48-a965-36d8cf6a8fa8 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "b29a8043-2006-4255-8036-669d2be594d3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.815297] env[63202]: DEBUG oslo_concurrency.lockutils [None req-610d9e96-076b-4d48-a965-36d8cf6a8fa8 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "b29a8043-2006-4255-8036-669d2be594d3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.815500] env[63202]: DEBUG oslo_concurrency.lockutils [None req-610d9e96-076b-4d48-a965-36d8cf6a8fa8 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "b29a8043-2006-4255-8036-669d2be594d3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.817749] env[63202]: INFO nova.compute.manager [None req-610d9e96-076b-4d48-a965-36d8cf6a8fa8 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: b29a8043-2006-4255-8036-669d2be594d3] Terminating instance [ 1021.819596] env[63202]: DEBUG nova.compute.manager [None req-610d9e96-076b-4d48-a965-36d8cf6a8fa8 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: b29a8043-2006-4255-8036-669d2be594d3] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1021.819797] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-610d9e96-076b-4d48-a965-36d8cf6a8fa8 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: b29a8043-2006-4255-8036-669d2be594d3] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1021.820633] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23cb39f3-087c-4bb4-8b0f-0dce4cccad23 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.828797] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-610d9e96-076b-4d48-a965-36d8cf6a8fa8 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: b29a8043-2006-4255-8036-669d2be594d3] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1021.829088] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-aa2f6ebf-3797-4966-a620-d13b1c73fe9d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.836073] env[63202]: DEBUG oslo_vmware.api [None req-610d9e96-076b-4d48-a965-36d8cf6a8fa8 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 1021.836073] env[63202]: value = "task-1385767" [ 1021.836073] env[63202]: _type = "Task" [ 1021.836073] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.843925] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3563b9ec-752c-4f6d-b5c5-e46a3a94bf81 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "dd146c71-c391-41e8-8cc7-7276f4a518d9" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.663s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.845151] env[63202]: DEBUG oslo_vmware.api [None req-610d9e96-076b-4d48-a965-36d8cf6a8fa8 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385767, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.896924] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Creating linked-clone VM from snapshot {{(pid=63202) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1021.898579] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-e2cec107-b0cb-400c-97b6-8cb204fa4f88 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.911221] env[63202]: DEBUG oslo_vmware.api [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1021.911221] env[63202]: value = "task-1385768" [ 1021.911221] env[63202]: _type = "Task" [ 1021.911221] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.924289] env[63202]: DEBUG oslo_vmware.api [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385768, 'name': CloneVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.073106] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Releasing lock "refresh_cache-3029c06f-4687-46f7-958b-831e4d0b241b" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.073425] env[63202]: DEBUG nova.compute.manager [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Instance network_info: |[{"id": "44f7adb9-2f63-4a9a-a041-0a0ae57c78c8", "address": "fa:16:3e:88:33:11", "network": {"id": "69295efd-3c50-45e5-bc09-4f0cef6b8f0e", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1062402518-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f651d18a9559499da27f301ee75bb2fb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dced2f3d-7fd3-4a42-836d-9f02dab4c949", "external-id": "nsx-vlan-transportzone-117", "segmentation_id": 117, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44f7adb9-2f", "ovs_interfaceid": "44f7adb9-2f63-4a9a-a041-0a0ae57c78c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1022.073969] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:88:33:11', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dced2f3d-7fd3-4a42-836d-9f02dab4c949', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '44f7adb9-2f63-4a9a-a041-0a0ae57c78c8', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1022.082874] env[63202]: DEBUG oslo.service.loopingcall [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1022.083162] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1022.083399] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-989c146e-78ec-4099-a158-1b06e3296253 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.104570] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1022.104570] env[63202]: value = "task-1385769" [ 1022.104570] env[63202]: _type = "Task" [ 1022.104570] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.113747] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385769, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.322732] env[63202]: DEBUG nova.compute.manager [req-9b95b2a7-0910-44dd-9cbf-21cf78306fcd req-ff91607d-c57a-4ac6-93ad-ae16b94ebe68 service nova] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Received event network-changed-44f7adb9-2f63-4a9a-a041-0a0ae57c78c8 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1022.322861] env[63202]: DEBUG nova.compute.manager [req-9b95b2a7-0910-44dd-9cbf-21cf78306fcd req-ff91607d-c57a-4ac6-93ad-ae16b94ebe68 service nova] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Refreshing instance network info cache due to event network-changed-44f7adb9-2f63-4a9a-a041-0a0ae57c78c8. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1022.323484] env[63202]: DEBUG oslo_concurrency.lockutils [req-9b95b2a7-0910-44dd-9cbf-21cf78306fcd req-ff91607d-c57a-4ac6-93ad-ae16b94ebe68 service nova] Acquiring lock "refresh_cache-3029c06f-4687-46f7-958b-831e4d0b241b" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1022.323484] env[63202]: DEBUG oslo_concurrency.lockutils [req-9b95b2a7-0910-44dd-9cbf-21cf78306fcd req-ff91607d-c57a-4ac6-93ad-ae16b94ebe68 service nova] Acquired lock "refresh_cache-3029c06f-4687-46f7-958b-831e4d0b241b" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.323484] env[63202]: DEBUG nova.network.neutron [req-9b95b2a7-0910-44dd-9cbf-21cf78306fcd req-ff91607d-c57a-4ac6-93ad-ae16b94ebe68 service nova] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Refreshing network info cache for port 44f7adb9-2f63-4a9a-a041-0a0ae57c78c8 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1022.344386] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1410fba-9e8b-4e09-94af-5b1239d8d66c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.350153] env[63202]: DEBUG oslo_vmware.api [None req-610d9e96-076b-4d48-a965-36d8cf6a8fa8 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385767, 'name': PowerOffVM_Task, 'duration_secs': 0.490318} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.350995] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-610d9e96-076b-4d48-a965-36d8cf6a8fa8 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: b29a8043-2006-4255-8036-669d2be594d3] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1022.351192] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-610d9e96-076b-4d48-a965-36d8cf6a8fa8 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: b29a8043-2006-4255-8036-669d2be594d3] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1022.351441] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5b004839-9bbd-40c4-8884-1a8ffb934794 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.355607] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1604ffd-cd96-4f7c-a81b-b316cd509f91 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.387715] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32dcba27-142b-4aa8-8bfc-200d481cedab {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.395777] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-114e6db8-2735-4070-a036-276b3491a851 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.411444] env[63202]: DEBUG oslo_concurrency.lockutils [None req-26d1a283-a645-4287-a46d-329c0a235176 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "dd146c71-c391-41e8-8cc7-7276f4a518d9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.411698] env[63202]: DEBUG oslo_concurrency.lockutils [None req-26d1a283-a645-4287-a46d-329c0a235176 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "dd146c71-c391-41e8-8cc7-7276f4a518d9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.411921] env[63202]: DEBUG oslo_concurrency.lockutils [None req-26d1a283-a645-4287-a46d-329c0a235176 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "dd146c71-c391-41e8-8cc7-7276f4a518d9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.412124] env[63202]: DEBUG oslo_concurrency.lockutils [None req-26d1a283-a645-4287-a46d-329c0a235176 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "dd146c71-c391-41e8-8cc7-7276f4a518d9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.412299] env[63202]: DEBUG oslo_concurrency.lockutils [None req-26d1a283-a645-4287-a46d-329c0a235176 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "dd146c71-c391-41e8-8cc7-7276f4a518d9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.414204] env[63202]: DEBUG nova.compute.provider_tree [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1022.415740] env[63202]: INFO nova.compute.manager [None req-26d1a283-a645-4287-a46d-329c0a235176 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Terminating instance [ 1022.420476] env[63202]: DEBUG nova.compute.manager [None req-26d1a283-a645-4287-a46d-329c0a235176 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1022.420686] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-26d1a283-a645-4287-a46d-329c0a235176 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1022.421700] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02d54680-2cd4-45bd-ba5a-ecb031c86ac4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.430832] env[63202]: DEBUG oslo_vmware.api [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385768, 'name': CloneVM_Task} progress is 94%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.433372] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-26d1a283-a645-4287-a46d-329c0a235176 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1022.433691] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-162ddad5-ea1a-4f8e-8192-d3c2a3269e7e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.440635] env[63202]: DEBUG oslo_vmware.api [None req-26d1a283-a645-4287-a46d-329c0a235176 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for the task: (returnval){ [ 1022.440635] env[63202]: value = "task-1385771" [ 1022.440635] env[63202]: _type = "Task" [ 1022.440635] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.450383] env[63202]: DEBUG oslo_vmware.api [None req-26d1a283-a645-4287-a46d-329c0a235176 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385771, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.615588] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385769, 'name': CreateVM_Task, 'duration_secs': 0.406476} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.616034] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1022.616542] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1022.616727] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.617105] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1022.617431] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5cb8eb84-ab52-4911-8541-853cb1e98e7a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.627981] env[63202]: DEBUG oslo_vmware.api [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Waiting for the task: (returnval){ [ 1022.627981] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52e528f3-6c40-60c9-ceb7-2bf44358b038" [ 1022.627981] env[63202]: _type = "Task" [ 1022.627981] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.636481] env[63202]: DEBUG oslo_vmware.api [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52e528f3-6c40-60c9-ceb7-2bf44358b038, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.852930] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-610d9e96-076b-4d48-a965-36d8cf6a8fa8 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: b29a8043-2006-4255-8036-669d2be594d3] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1022.853284] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-610d9e96-076b-4d48-a965-36d8cf6a8fa8 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: b29a8043-2006-4255-8036-669d2be594d3] Deleting contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1022.853572] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-610d9e96-076b-4d48-a965-36d8cf6a8fa8 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Deleting the datastore file [datastore1] b29a8043-2006-4255-8036-669d2be594d3 {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1022.853858] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b1a6a8be-e6f3-40b0-9960-995ce8bfb275 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.862399] env[63202]: DEBUG oslo_vmware.api [None req-610d9e96-076b-4d48-a965-36d8cf6a8fa8 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 1022.862399] env[63202]: value = "task-1385772" [ 1022.862399] env[63202]: _type = "Task" [ 1022.862399] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.869808] env[63202]: DEBUG oslo_concurrency.lockutils [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "c2b6f3a4-7414-4fc4-893b-9b613fc5e381" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.870091] env[63202]: DEBUG oslo_concurrency.lockutils [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "c2b6f3a4-7414-4fc4-893b-9b613fc5e381" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.874565] env[63202]: DEBUG oslo_vmware.api [None req-610d9e96-076b-4d48-a965-36d8cf6a8fa8 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385772, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.919063] env[63202]: DEBUG nova.scheduler.client.report [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1022.940285] env[63202]: DEBUG oslo_vmware.api [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385768, 'name': CloneVM_Task} progress is 100%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.953978] env[63202]: DEBUG oslo_vmware.api [None req-26d1a283-a645-4287-a46d-329c0a235176 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385771, 'name': PowerOffVM_Task, 'duration_secs': 0.200794} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.954263] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-26d1a283-a645-4287-a46d-329c0a235176 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1022.954437] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-26d1a283-a645-4287-a46d-329c0a235176 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1022.954689] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a3044c34-2637-4e56-8527-8ce3d44df1c2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.027994] env[63202]: DEBUG nova.network.neutron [req-9b95b2a7-0910-44dd-9cbf-21cf78306fcd req-ff91607d-c57a-4ac6-93ad-ae16b94ebe68 service nova] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Updated VIF entry in instance network info cache for port 44f7adb9-2f63-4a9a-a041-0a0ae57c78c8. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1023.028425] env[63202]: DEBUG nova.network.neutron [req-9b95b2a7-0910-44dd-9cbf-21cf78306fcd req-ff91607d-c57a-4ac6-93ad-ae16b94ebe68 service nova] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Updating instance_info_cache with network_info: [{"id": "44f7adb9-2f63-4a9a-a041-0a0ae57c78c8", "address": "fa:16:3e:88:33:11", "network": {"id": "69295efd-3c50-45e5-bc09-4f0cef6b8f0e", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1062402518-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f651d18a9559499da27f301ee75bb2fb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dced2f3d-7fd3-4a42-836d-9f02dab4c949", "external-id": "nsx-vlan-transportzone-117", "segmentation_id": 117, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44f7adb9-2f", "ovs_interfaceid": "44f7adb9-2f63-4a9a-a041-0a0ae57c78c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.033159] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-26d1a283-a645-4287-a46d-329c0a235176 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1023.033370] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-26d1a283-a645-4287-a46d-329c0a235176 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Deleting contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1023.033557] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-26d1a283-a645-4287-a46d-329c0a235176 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Deleting the datastore file [datastore1] dd146c71-c391-41e8-8cc7-7276f4a518d9 {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1023.033821] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c325c9d2-c703-4eb3-a19c-24a9f6994df6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.040981] env[63202]: DEBUG oslo_vmware.api [None req-26d1a283-a645-4287-a46d-329c0a235176 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for the task: (returnval){ [ 1023.040981] env[63202]: value = "task-1385774" [ 1023.040981] env[63202]: _type = "Task" [ 1023.040981] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.049542] env[63202]: DEBUG oslo_vmware.api [None req-26d1a283-a645-4287-a46d-329c0a235176 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385774, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.138651] env[63202]: DEBUG oslo_vmware.api [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52e528f3-6c40-60c9-ceb7-2bf44358b038, 'name': SearchDatastore_Task, 'duration_secs': 0.01754} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.139075] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1023.139342] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1023.139614] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1023.139859] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.140139] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1023.140421] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a55ccf38-036e-4ac6-9d66-7d1fc7d23c51 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.148194] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1023.148409] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1023.149176] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5538a04a-47a3-4fc1-a0f7-b704d5fe81c4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.154494] env[63202]: DEBUG oslo_vmware.api [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Waiting for the task: (returnval){ [ 1023.154494] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52e22d5d-8b75-9dec-d82e-7d501b3fc5a1" [ 1023.154494] env[63202]: _type = "Task" [ 1023.154494] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.164496] env[63202]: DEBUG oslo_vmware.api [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52e22d5d-8b75-9dec-d82e-7d501b3fc5a1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.372349] env[63202]: DEBUG oslo_vmware.api [None req-610d9e96-076b-4d48-a965-36d8cf6a8fa8 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385772, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.180994} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.372608] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-610d9e96-076b-4d48-a965-36d8cf6a8fa8 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1023.372794] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-610d9e96-076b-4d48-a965-36d8cf6a8fa8 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: b29a8043-2006-4255-8036-669d2be594d3] Deleted contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1023.372974] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-610d9e96-076b-4d48-a965-36d8cf6a8fa8 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: b29a8043-2006-4255-8036-669d2be594d3] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1023.373166] env[63202]: INFO nova.compute.manager [None req-610d9e96-076b-4d48-a965-36d8cf6a8fa8 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: b29a8043-2006-4255-8036-669d2be594d3] Took 1.55 seconds to destroy the instance on the hypervisor. [ 1023.373407] env[63202]: DEBUG oslo.service.loopingcall [None req-610d9e96-076b-4d48-a965-36d8cf6a8fa8 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1023.373600] env[63202]: DEBUG nova.compute.manager [-] [instance: b29a8043-2006-4255-8036-669d2be594d3] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1023.373692] env[63202]: DEBUG nova.network.neutron [-] [instance: b29a8043-2006-4255-8036-669d2be594d3] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1023.375465] env[63202]: DEBUG nova.compute.manager [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1023.427762] env[63202]: DEBUG oslo_vmware.api [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385768, 'name': CloneVM_Task, 'duration_secs': 1.093213} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.428047] env[63202]: INFO nova.virt.vmwareapi.vmops [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Created linked-clone VM from snapshot [ 1023.428871] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04879368-3f4e-4870-a01e-ed696b05dfc5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.431746] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.236s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.432240] env[63202]: DEBUG nova.compute.manager [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1023.440692] env[63202]: DEBUG nova.virt.vmwareapi.images [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Uploading image 3db9d5e2-204c-47e7-ae45-3ed10283296d {{(pid=63202) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1023.470148] env[63202]: DEBUG oslo_vmware.rw_handles [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1023.470148] env[63202]: value = "vm-294204" [ 1023.470148] env[63202]: _type = "VirtualMachine" [ 1023.470148] env[63202]: }. {{(pid=63202) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1023.470571] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-06cb9ab2-c3bb-4d7c-93c4-f4f52056e832 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.480046] env[63202]: DEBUG oslo_vmware.rw_handles [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lease: (returnval){ [ 1023.480046] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52182c4c-dbd8-a850-f7fb-ab1fe36f7219" [ 1023.480046] env[63202]: _type = "HttpNfcLease" [ 1023.480046] env[63202]: } obtained for exporting VM: (result){ [ 1023.480046] env[63202]: value = "vm-294204" [ 1023.480046] env[63202]: _type = "VirtualMachine" [ 1023.480046] env[63202]: }. {{(pid=63202) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1023.480357] env[63202]: DEBUG oslo_vmware.api [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the lease: (returnval){ [ 1023.480357] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52182c4c-dbd8-a850-f7fb-ab1fe36f7219" [ 1023.480357] env[63202]: _type = "HttpNfcLease" [ 1023.480357] env[63202]: } to be ready. {{(pid=63202) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1023.489669] env[63202]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1023.489669] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52182c4c-dbd8-a850-f7fb-ab1fe36f7219" [ 1023.489669] env[63202]: _type = "HttpNfcLease" [ 1023.489669] env[63202]: } is initializing. {{(pid=63202) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1023.531535] env[63202]: DEBUG oslo_concurrency.lockutils [req-9b95b2a7-0910-44dd-9cbf-21cf78306fcd req-ff91607d-c57a-4ac6-93ad-ae16b94ebe68 service nova] Releasing lock "refresh_cache-3029c06f-4687-46f7-958b-831e4d0b241b" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1023.550643] env[63202]: DEBUG oslo_vmware.api [None req-26d1a283-a645-4287-a46d-329c0a235176 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385774, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140854} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.550922] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-26d1a283-a645-4287-a46d-329c0a235176 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1023.551130] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-26d1a283-a645-4287-a46d-329c0a235176 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Deleted contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1023.551311] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-26d1a283-a645-4287-a46d-329c0a235176 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1023.551483] env[63202]: INFO nova.compute.manager [None req-26d1a283-a645-4287-a46d-329c0a235176 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1023.551719] env[63202]: DEBUG oslo.service.loopingcall [None req-26d1a283-a645-4287-a46d-329c0a235176 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1023.551937] env[63202]: DEBUG nova.compute.manager [-] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1023.552081] env[63202]: DEBUG nova.network.neutron [-] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1023.666203] env[63202]: DEBUG oslo_vmware.api [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52e22d5d-8b75-9dec-d82e-7d501b3fc5a1, 'name': SearchDatastore_Task, 'duration_secs': 0.008985} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.666704] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e1e77084-f7c3-4ebb-94d5-b23bafc6661e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.672086] env[63202]: DEBUG oslo_vmware.api [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Waiting for the task: (returnval){ [ 1023.672086] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52755a7c-9a45-8d7d-af73-f75aae9f7906" [ 1023.672086] env[63202]: _type = "Task" [ 1023.672086] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.679778] env[63202]: DEBUG oslo_vmware.api [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52755a7c-9a45-8d7d-af73-f75aae9f7906, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.799404] env[63202]: DEBUG nova.compute.manager [req-8f595921-ade4-4575-827a-3e9c0f055146 req-27a32804-1a92-4138-9e80-17600735a381 service nova] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Received event network-vif-deleted-0bd2810b-f01e-4cd8-9c3e-76ac2339047a {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1023.799574] env[63202]: INFO nova.compute.manager [req-8f595921-ade4-4575-827a-3e9c0f055146 req-27a32804-1a92-4138-9e80-17600735a381 service nova] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Neutron deleted interface 0bd2810b-f01e-4cd8-9c3e-76ac2339047a; detaching it from the instance and deleting it from the info cache [ 1023.799777] env[63202]: DEBUG nova.network.neutron [req-8f595921-ade4-4575-827a-3e9c0f055146 req-27a32804-1a92-4138-9e80-17600735a381 service nova] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.897705] env[63202]: DEBUG oslo_concurrency.lockutils [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.897967] env[63202]: DEBUG oslo_concurrency.lockutils [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.899602] env[63202]: INFO nova.compute.claims [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1023.937273] env[63202]: DEBUG nova.compute.utils [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1023.939523] env[63202]: DEBUG nova.compute.manager [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1023.939667] env[63202]: DEBUG nova.network.neutron [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1023.977821] env[63202]: DEBUG nova.policy [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b59e7e62d92e4f5eb0fee41e8f5196d7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f508ec04edc844a19640a8a85f27e5b9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 1023.987828] env[63202]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1023.987828] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52182c4c-dbd8-a850-f7fb-ab1fe36f7219" [ 1023.987828] env[63202]: _type = "HttpNfcLease" [ 1023.987828] env[63202]: } is ready. {{(pid=63202) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1023.988145] env[63202]: DEBUG oslo_vmware.rw_handles [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1023.988145] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52182c4c-dbd8-a850-f7fb-ab1fe36f7219" [ 1023.988145] env[63202]: _type = "HttpNfcLease" [ 1023.988145] env[63202]: }. {{(pid=63202) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1023.988861] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55d398ea-2701-4546-ad68-72d9ed4882e3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.997235] env[63202]: DEBUG oslo_vmware.rw_handles [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b09717-c9a4-f5dc-9b54-1c5cd758adb9/disk-0.vmdk from lease info. {{(pid=63202) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1023.997408] env[63202]: DEBUG oslo_vmware.rw_handles [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b09717-c9a4-f5dc-9b54-1c5cd758adb9/disk-0.vmdk for reading. {{(pid=63202) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1024.086474] env[63202]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-f25ac863-c201-4c45-a997-b015c3ffe6f2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.127467] env[63202]: DEBUG nova.network.neutron [-] [instance: b29a8043-2006-4255-8036-669d2be594d3] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1024.183484] env[63202]: DEBUG oslo_vmware.api [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52755a7c-9a45-8d7d-af73-f75aae9f7906, 'name': SearchDatastore_Task, 'duration_secs': 0.016303} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.183484] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.183484] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] 3029c06f-4687-46f7-958b-831e4d0b241b/3029c06f-4687-46f7-958b-831e4d0b241b.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1024.183769] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-426d35c0-e07b-4dbe-b27b-43c2ba483d5e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.191748] env[63202]: DEBUG oslo_vmware.api [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Waiting for the task: (returnval){ [ 1024.191748] env[63202]: value = "task-1385776" [ 1024.191748] env[63202]: _type = "Task" [ 1024.191748] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.199854] env[63202]: DEBUG oslo_vmware.api [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385776, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.279211] env[63202]: DEBUG nova.network.neutron [-] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1024.303423] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-172ade6a-1d7f-4c08-9553-ec0df0b8978b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.313596] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-075339e8-1ced-49dd-b830-0dea125267af {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.344101] env[63202]: DEBUG nova.compute.manager [req-8f595921-ade4-4575-827a-3e9c0f055146 req-27a32804-1a92-4138-9e80-17600735a381 service nova] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Detach interface failed, port_id=0bd2810b-f01e-4cd8-9c3e-76ac2339047a, reason: Instance dd146c71-c391-41e8-8cc7-7276f4a518d9 could not be found. {{(pid=63202) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1024.357067] env[63202]: DEBUG nova.compute.manager [req-387d7a4f-5e86-4c1c-9953-bfe7e5875b49 req-57f04fd4-06bd-4935-afe8-55e4a06d2744 service nova] [instance: b29a8043-2006-4255-8036-669d2be594d3] Received event network-vif-deleted-cefc0ade-cf96-4e99-8ebf-a64ae8190649 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1024.442696] env[63202]: DEBUG nova.compute.manager [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1024.630692] env[63202]: INFO nova.compute.manager [-] [instance: b29a8043-2006-4255-8036-669d2be594d3] Took 1.26 seconds to deallocate network for instance. [ 1024.685109] env[63202]: DEBUG nova.network.neutron [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Successfully created port: 5f6faad3-ee63-4318-abef-e6442d090681 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1024.704537] env[63202]: DEBUG oslo_vmware.api [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385776, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.468507} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.705911] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] 3029c06f-4687-46f7-958b-831e4d0b241b/3029c06f-4687-46f7-958b-831e4d0b241b.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1024.705911] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1024.706990] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dbaae625-8345-47e8-9b23-64887c02434c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.714433] env[63202]: DEBUG oslo_vmware.api [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Waiting for the task: (returnval){ [ 1024.714433] env[63202]: value = "task-1385777" [ 1024.714433] env[63202]: _type = "Task" [ 1024.714433] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.723553] env[63202]: DEBUG oslo_vmware.api [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385777, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.783391] env[63202]: INFO nova.compute.manager [-] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Took 1.23 seconds to deallocate network for instance. [ 1025.081982] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57eba0dd-cf6d-4409-aca6-52f2d011e887 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.093234] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10b49b46-5631-4b7e-b554-79a6de28d040 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.127837] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0ef17f9-5ab5-484f-b640-2744370645ae {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.137469] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6fca0a6-7764-4136-b50e-a91cc97e503c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.145694] env[63202]: DEBUG oslo_concurrency.lockutils [None req-610d9e96-076b-4d48-a965-36d8cf6a8fa8 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.157672] env[63202]: DEBUG nova.compute.provider_tree [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1025.224524] env[63202]: DEBUG oslo_vmware.api [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385777, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066349} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.225243] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1025.225954] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-231686f2-2656-4a22-a58f-f83d709f84c5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.250660] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Reconfiguring VM instance instance-0000005e to attach disk [datastore1] 3029c06f-4687-46f7-958b-831e4d0b241b/3029c06f-4687-46f7-958b-831e4d0b241b.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1025.251773] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c5058561-86ac-4f9c-bb77-ce7d921aae7d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.271895] env[63202]: DEBUG oslo_vmware.api [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Waiting for the task: (returnval){ [ 1025.271895] env[63202]: value = "task-1385778" [ 1025.271895] env[63202]: _type = "Task" [ 1025.271895] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.281455] env[63202]: DEBUG oslo_vmware.api [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385778, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.294117] env[63202]: DEBUG oslo_concurrency.lockutils [None req-26d1a283-a645-4287-a46d-329c0a235176 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.454436] env[63202]: DEBUG nova.compute.manager [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1025.476804] env[63202]: DEBUG nova.virt.hardware [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1025.476950] env[63202]: DEBUG nova.virt.hardware [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1025.477163] env[63202]: DEBUG nova.virt.hardware [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1025.477384] env[63202]: DEBUG nova.virt.hardware [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1025.478253] env[63202]: DEBUG nova.virt.hardware [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1025.478253] env[63202]: DEBUG nova.virt.hardware [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1025.478424] env[63202]: DEBUG nova.virt.hardware [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1025.478844] env[63202]: DEBUG nova.virt.hardware [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1025.478844] env[63202]: DEBUG nova.virt.hardware [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1025.479690] env[63202]: DEBUG nova.virt.hardware [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1025.479690] env[63202]: DEBUG nova.virt.hardware [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1025.480765] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4a3213a-bde4-4364-9249-32fb69cdf3c7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.489862] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fafa83dd-b2f2-453a-9f63-393f8fbf9fcc {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.593636] env[63202]: DEBUG oslo_concurrency.lockutils [None req-41f6cd42-ad0d-4d98-b912-a0197bfe098f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Acquiring lock "0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.594012] env[63202]: DEBUG oslo_concurrency.lockutils [None req-41f6cd42-ad0d-4d98-b912-a0197bfe098f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lock "0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.594289] env[63202]: DEBUG oslo_concurrency.lockutils [None req-41f6cd42-ad0d-4d98-b912-a0197bfe098f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Acquiring lock "0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.594544] env[63202]: DEBUG oslo_concurrency.lockutils [None req-41f6cd42-ad0d-4d98-b912-a0197bfe098f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lock "0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.594928] env[63202]: DEBUG oslo_concurrency.lockutils [None req-41f6cd42-ad0d-4d98-b912-a0197bfe098f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lock "0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.597056] env[63202]: INFO nova.compute.manager [None req-41f6cd42-ad0d-4d98-b912-a0197bfe098f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Terminating instance [ 1025.599187] env[63202]: DEBUG nova.compute.manager [None req-41f6cd42-ad0d-4d98-b912-a0197bfe098f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1025.599391] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-41f6cd42-ad0d-4d98-b912-a0197bfe098f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1025.600248] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b01c7c4d-da68-4787-8077-75086dc6d2df {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.608053] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-41f6cd42-ad0d-4d98-b912-a0197bfe098f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1025.609076] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2e6e3496-0f1c-46ff-b3f4-68073b6e6346 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.614595] env[63202]: DEBUG oslo_vmware.api [None req-41f6cd42-ad0d-4d98-b912-a0197bfe098f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Waiting for the task: (returnval){ [ 1025.614595] env[63202]: value = "task-1385779" [ 1025.614595] env[63202]: _type = "Task" [ 1025.614595] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.629178] env[63202]: DEBUG oslo_vmware.api [None req-41f6cd42-ad0d-4d98-b912-a0197bfe098f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385779, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.661446] env[63202]: DEBUG nova.scheduler.client.report [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1025.786688] env[63202]: DEBUG oslo_vmware.api [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385778, 'name': ReconfigVM_Task, 'duration_secs': 0.265284} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.787190] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Reconfigured VM instance instance-0000005e to attach disk [datastore1] 3029c06f-4687-46f7-958b-831e4d0b241b/3029c06f-4687-46f7-958b-831e4d0b241b.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1025.788291] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0732048d-d81b-4aad-9a93-3bb17ec19f0d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.797368] env[63202]: DEBUG oslo_vmware.api [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Waiting for the task: (returnval){ [ 1025.797368] env[63202]: value = "task-1385780" [ 1025.797368] env[63202]: _type = "Task" [ 1025.797368] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.810800] env[63202]: DEBUG oslo_vmware.api [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385780, 'name': Rename_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.124547] env[63202]: DEBUG oslo_vmware.api [None req-41f6cd42-ad0d-4d98-b912-a0197bfe098f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385779, 'name': PowerOffVM_Task, 'duration_secs': 0.209754} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.124895] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-41f6cd42-ad0d-4d98-b912-a0197bfe098f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1026.125058] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-41f6cd42-ad0d-4d98-b912-a0197bfe098f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1026.125345] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4e0b1d07-98e5-4b54-a88a-3070544dc25e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.166624] env[63202]: DEBUG oslo_concurrency.lockutils [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.268s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.167209] env[63202]: DEBUG nova.compute.manager [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1026.170533] env[63202]: DEBUG oslo_concurrency.lockutils [None req-610d9e96-076b-4d48-a965-36d8cf6a8fa8 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.025s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.170822] env[63202]: DEBUG nova.objects.instance [None req-610d9e96-076b-4d48-a965-36d8cf6a8fa8 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lazy-loading 'resources' on Instance uuid b29a8043-2006-4255-8036-669d2be594d3 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1026.188266] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-41f6cd42-ad0d-4d98-b912-a0197bfe098f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1026.188602] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-41f6cd42-ad0d-4d98-b912-a0197bfe098f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Deleting contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1026.188721] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-41f6cd42-ad0d-4d98-b912-a0197bfe098f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Deleting the datastore file [datastore1] 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1026.188987] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-426e51ac-3a86-451e-aaca-340985499b9d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.196262] env[63202]: DEBUG oslo_vmware.api [None req-41f6cd42-ad0d-4d98-b912-a0197bfe098f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Waiting for the task: (returnval){ [ 1026.196262] env[63202]: value = "task-1385782" [ 1026.196262] env[63202]: _type = "Task" [ 1026.196262] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.205209] env[63202]: DEBUG oslo_vmware.api [None req-41f6cd42-ad0d-4d98-b912-a0197bfe098f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385782, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.238573] env[63202]: DEBUG nova.compute.manager [req-9e2a7cd6-f56c-488a-b11c-f010ffe04701 req-9f215d72-0e6d-4956-af20-9ee5f58ba154 service nova] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Received event network-vif-plugged-5f6faad3-ee63-4318-abef-e6442d090681 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1026.238829] env[63202]: DEBUG oslo_concurrency.lockutils [req-9e2a7cd6-f56c-488a-b11c-f010ffe04701 req-9f215d72-0e6d-4956-af20-9ee5f58ba154 service nova] Acquiring lock "7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.239075] env[63202]: DEBUG oslo_concurrency.lockutils [req-9e2a7cd6-f56c-488a-b11c-f010ffe04701 req-9f215d72-0e6d-4956-af20-9ee5f58ba154 service nova] Lock "7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.239268] env[63202]: DEBUG oslo_concurrency.lockutils [req-9e2a7cd6-f56c-488a-b11c-f010ffe04701 req-9f215d72-0e6d-4956-af20-9ee5f58ba154 service nova] Lock "7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.239491] env[63202]: DEBUG nova.compute.manager [req-9e2a7cd6-f56c-488a-b11c-f010ffe04701 req-9f215d72-0e6d-4956-af20-9ee5f58ba154 service nova] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] No waiting events found dispatching network-vif-plugged-5f6faad3-ee63-4318-abef-e6442d090681 {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1026.239621] env[63202]: WARNING nova.compute.manager [req-9e2a7cd6-f56c-488a-b11c-f010ffe04701 req-9f215d72-0e6d-4956-af20-9ee5f58ba154 service nova] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Received unexpected event network-vif-plugged-5f6faad3-ee63-4318-abef-e6442d090681 for instance with vm_state building and task_state spawning. [ 1026.309288] env[63202]: DEBUG oslo_vmware.api [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385780, 'name': Rename_Task, 'duration_secs': 0.141553} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.309650] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1026.310014] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-64f0475b-49ee-4808-83b5-f26124e70d78 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.316996] env[63202]: DEBUG oslo_vmware.api [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Waiting for the task: (returnval){ [ 1026.316996] env[63202]: value = "task-1385783" [ 1026.316996] env[63202]: _type = "Task" [ 1026.316996] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.331603] env[63202]: DEBUG oslo_vmware.api [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385783, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.382964] env[63202]: DEBUG nova.network.neutron [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Successfully updated port: 5f6faad3-ee63-4318-abef-e6442d090681 {{(pid=63202) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1026.674680] env[63202]: DEBUG nova.compute.utils [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1026.676738] env[63202]: DEBUG nova.compute.manager [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1026.677022] env[63202]: DEBUG nova.network.neutron [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1026.708577] env[63202]: DEBUG oslo_vmware.api [None req-41f6cd42-ad0d-4d98-b912-a0197bfe098f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385782, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.23873} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.709069] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-41f6cd42-ad0d-4d98-b912-a0197bfe098f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1026.709069] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-41f6cd42-ad0d-4d98-b912-a0197bfe098f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Deleted contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1026.709251] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-41f6cd42-ad0d-4d98-b912-a0197bfe098f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1026.709300] env[63202]: INFO nova.compute.manager [None req-41f6cd42-ad0d-4d98-b912-a0197bfe098f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1026.709522] env[63202]: DEBUG oslo.service.loopingcall [None req-41f6cd42-ad0d-4d98-b912-a0197bfe098f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1026.709728] env[63202]: DEBUG nova.compute.manager [-] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1026.709821] env[63202]: DEBUG nova.network.neutron [-] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1026.725832] env[63202]: DEBUG nova.policy [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1236175893a3401cb01b04604168f1e0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '22352c25bca8416a948014391a5389ac', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 1026.829545] env[63202]: DEBUG oslo_vmware.api [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385783, 'name': PowerOnVM_Task, 'duration_secs': 0.492697} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.830086] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1026.830384] env[63202]: INFO nova.compute.manager [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Took 6.70 seconds to spawn the instance on the hypervisor. [ 1026.830683] env[63202]: DEBUG nova.compute.manager [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1026.831590] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baae33a5-3f2a-429c-9595-aa7d7eb72030 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.842272] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f8eae3e-b2a8-4b1e-97a2-745830c9d5d1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.850768] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-130348b3-0f7b-4338-bcb9-4ce6240d6a12 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.883172] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcddfe5a-6642-4c1b-842e-4444cb11744f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.888270] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "refresh_cache-7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1026.888270] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquired lock "refresh_cache-7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.888455] env[63202]: DEBUG nova.network.neutron [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1026.893792] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-190d9a5a-46d6-41a3-92b9-dcfd0e8a40b2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.911989] env[63202]: DEBUG nova.compute.provider_tree [None req-610d9e96-076b-4d48-a965-36d8cf6a8fa8 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1027.076030] env[63202]: DEBUG nova.network.neutron [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Successfully created port: ea5e41d2-19a6-4899-9bfa-c9b18de354e6 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1027.177887] env[63202]: DEBUG nova.compute.manager [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1027.307680] env[63202]: DEBUG nova.compute.manager [req-7a1b5c60-7292-4cc8-bdc2-322366eda43d req-c47607b0-0f0e-4f55-a812-ae6a3111c93a service nova] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Received event network-vif-deleted-e5621970-d0a8-4803-94ef-f6d577bdf6d2 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1027.308092] env[63202]: INFO nova.compute.manager [req-7a1b5c60-7292-4cc8-bdc2-322366eda43d req-c47607b0-0f0e-4f55-a812-ae6a3111c93a service nova] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Neutron deleted interface e5621970-d0a8-4803-94ef-f6d577bdf6d2; detaching it from the instance and deleting it from the info cache [ 1027.308092] env[63202]: DEBUG nova.network.neutron [req-7a1b5c60-7292-4cc8-bdc2-322366eda43d req-c47607b0-0f0e-4f55-a812-ae6a3111c93a service nova] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.353523] env[63202]: INFO nova.compute.manager [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Took 13.24 seconds to build instance. [ 1027.416443] env[63202]: DEBUG nova.scheduler.client.report [None req-610d9e96-076b-4d48-a965-36d8cf6a8fa8 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1027.428889] env[63202]: DEBUG nova.network.neutron [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1027.563147] env[63202]: DEBUG nova.network.neutron [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Updating instance_info_cache with network_info: [{"id": "5f6faad3-ee63-4318-abef-e6442d090681", "address": "fa:16:3e:88:4a:c0", "network": {"id": "5a3b4c9b-2ca7-4f8b-8bbb-ea000db91402", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-543677750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f508ec04edc844a19640a8a85f27e5b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f6faad3-ee", "ovs_interfaceid": "5f6faad3-ee63-4318-abef-e6442d090681", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.602609] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bd8645bc-e2f4-4a6b-a7b2-75b30ea9aa23 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Acquiring lock "3029c06f-4687-46f7-958b-831e4d0b241b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.668205] env[63202]: DEBUG nova.network.neutron [-] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.813339] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ce25488e-1a7b-409c-86fd-c3c9fbcb5645 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.824132] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb181308-aa6f-489c-8f9a-4f9e0fc1c5bf {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.857902] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b08a8426-4ea6-4d26-86d0-16f1f3f9e7aa tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Lock "3029c06f-4687-46f7-958b-831e4d0b241b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.749s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.858336] env[63202]: DEBUG nova.compute.manager [req-7a1b5c60-7292-4cc8-bdc2-322366eda43d req-c47607b0-0f0e-4f55-a812-ae6a3111c93a service nova] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Detach interface failed, port_id=e5621970-d0a8-4803-94ef-f6d577bdf6d2, reason: Instance 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb could not be found. {{(pid=63202) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1027.858728] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bd8645bc-e2f4-4a6b-a7b2-75b30ea9aa23 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Lock "3029c06f-4687-46f7-958b-831e4d0b241b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.256s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.858921] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bd8645bc-e2f4-4a6b-a7b2-75b30ea9aa23 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Acquiring lock "3029c06f-4687-46f7-958b-831e4d0b241b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.859172] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bd8645bc-e2f4-4a6b-a7b2-75b30ea9aa23 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Lock "3029c06f-4687-46f7-958b-831e4d0b241b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.859348] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bd8645bc-e2f4-4a6b-a7b2-75b30ea9aa23 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Lock "3029c06f-4687-46f7-958b-831e4d0b241b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.861745] env[63202]: INFO nova.compute.manager [None req-bd8645bc-e2f4-4a6b-a7b2-75b30ea9aa23 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Terminating instance [ 1027.863815] env[63202]: DEBUG nova.compute.manager [None req-bd8645bc-e2f4-4a6b-a7b2-75b30ea9aa23 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1027.865019] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-bd8645bc-e2f4-4a6b-a7b2-75b30ea9aa23 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1027.865019] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9b8b5b0-aa94-4d8f-a4c7-e11de73cddee {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.873228] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd8645bc-e2f4-4a6b-a7b2-75b30ea9aa23 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1027.873515] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d07277c1-65fc-47d9-8396-c2005dbf06c4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.881185] env[63202]: DEBUG oslo_vmware.api [None req-bd8645bc-e2f4-4a6b-a7b2-75b30ea9aa23 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Waiting for the task: (returnval){ [ 1027.881185] env[63202]: value = "task-1385784" [ 1027.881185] env[63202]: _type = "Task" [ 1027.881185] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.890537] env[63202]: DEBUG oslo_vmware.api [None req-bd8645bc-e2f4-4a6b-a7b2-75b30ea9aa23 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385784, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.923951] env[63202]: DEBUG oslo_concurrency.lockutils [None req-610d9e96-076b-4d48-a965-36d8cf6a8fa8 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.753s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.927301] env[63202]: DEBUG oslo_concurrency.lockutils [None req-26d1a283-a645-4287-a46d-329c0a235176 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.633s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.927301] env[63202]: DEBUG oslo_concurrency.lockutils [None req-26d1a283-a645-4287-a46d-329c0a235176 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.949095] env[63202]: INFO nova.scheduler.client.report [None req-610d9e96-076b-4d48-a965-36d8cf6a8fa8 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Deleted allocations for instance b29a8043-2006-4255-8036-669d2be594d3 [ 1027.951373] env[63202]: INFO nova.scheduler.client.report [None req-26d1a283-a645-4287-a46d-329c0a235176 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Deleted allocations for instance dd146c71-c391-41e8-8cc7-7276f4a518d9 [ 1028.065748] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Releasing lock "refresh_cache-7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.066254] env[63202]: DEBUG nova.compute.manager [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Instance network_info: |[{"id": "5f6faad3-ee63-4318-abef-e6442d090681", "address": "fa:16:3e:88:4a:c0", "network": {"id": "5a3b4c9b-2ca7-4f8b-8bbb-ea000db91402", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-543677750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f508ec04edc844a19640a8a85f27e5b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f6faad3-ee", "ovs_interfaceid": "5f6faad3-ee63-4318-abef-e6442d090681", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1028.066819] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:88:4a:c0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5f6faad3-ee63-4318-abef-e6442d090681', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1028.077032] env[63202]: DEBUG oslo.service.loopingcall [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1028.077032] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1028.077032] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c9dc828c-c768-4515-b276-ec3e0ed1b83a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.098259] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1028.098259] env[63202]: value = "task-1385785" [ 1028.098259] env[63202]: _type = "Task" [ 1028.098259] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.107377] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385785, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.171835] env[63202]: INFO nova.compute.manager [-] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Took 1.46 seconds to deallocate network for instance. [ 1028.188161] env[63202]: DEBUG nova.compute.manager [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1028.216113] env[63202]: DEBUG nova.virt.hardware [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1028.216446] env[63202]: DEBUG nova.virt.hardware [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1028.216607] env[63202]: DEBUG nova.virt.hardware [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1028.216866] env[63202]: DEBUG nova.virt.hardware [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1028.217071] env[63202]: DEBUG nova.virt.hardware [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1028.217379] env[63202]: DEBUG nova.virt.hardware [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1028.217506] env[63202]: DEBUG nova.virt.hardware [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1028.217731] env[63202]: DEBUG nova.virt.hardware [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1028.217960] env[63202]: DEBUG nova.virt.hardware [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1028.218171] env[63202]: DEBUG nova.virt.hardware [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1028.218353] env[63202]: DEBUG nova.virt.hardware [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1028.219530] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f252d383-ae88-433e-872e-e0bba920c532 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.228546] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-734b2068-0ec1-422d-9863-47079b1737f2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.268995] env[63202]: DEBUG nova.compute.manager [req-026dd47b-4fe1-4f12-b1ff-0949d85df426 req-221f1ad1-2961-48f2-9985-4110437bff90 service nova] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Received event network-changed-5f6faad3-ee63-4318-abef-e6442d090681 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1028.269297] env[63202]: DEBUG nova.compute.manager [req-026dd47b-4fe1-4f12-b1ff-0949d85df426 req-221f1ad1-2961-48f2-9985-4110437bff90 service nova] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Refreshing instance network info cache due to event network-changed-5f6faad3-ee63-4318-abef-e6442d090681. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1028.269603] env[63202]: DEBUG oslo_concurrency.lockutils [req-026dd47b-4fe1-4f12-b1ff-0949d85df426 req-221f1ad1-2961-48f2-9985-4110437bff90 service nova] Acquiring lock "refresh_cache-7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.269802] env[63202]: DEBUG oslo_concurrency.lockutils [req-026dd47b-4fe1-4f12-b1ff-0949d85df426 req-221f1ad1-2961-48f2-9985-4110437bff90 service nova] Acquired lock "refresh_cache-7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.270047] env[63202]: DEBUG nova.network.neutron [req-026dd47b-4fe1-4f12-b1ff-0949d85df426 req-221f1ad1-2961-48f2-9985-4110437bff90 service nova] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Refreshing network info cache for port 5f6faad3-ee63-4318-abef-e6442d090681 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1028.393442] env[63202]: DEBUG oslo_vmware.api [None req-bd8645bc-e2f4-4a6b-a7b2-75b30ea9aa23 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385784, 'name': PowerOffVM_Task, 'duration_secs': 0.235299} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.393642] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd8645bc-e2f4-4a6b-a7b2-75b30ea9aa23 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1028.393762] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-bd8645bc-e2f4-4a6b-a7b2-75b30ea9aa23 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1028.394041] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-98219663-1cc7-4d1b-bed6-e6925df370b7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.464138] env[63202]: DEBUG oslo_concurrency.lockutils [None req-610d9e96-076b-4d48-a965-36d8cf6a8fa8 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "b29a8043-2006-4255-8036-669d2be594d3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.649s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.466179] env[63202]: DEBUG oslo_concurrency.lockutils [None req-26d1a283-a645-4287-a46d-329c0a235176 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "dd146c71-c391-41e8-8cc7-7276f4a518d9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.054s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.478627] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-bd8645bc-e2f4-4a6b-a7b2-75b30ea9aa23 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1028.478939] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-bd8645bc-e2f4-4a6b-a7b2-75b30ea9aa23 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Deleting contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1028.479276] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd8645bc-e2f4-4a6b-a7b2-75b30ea9aa23 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Deleting the datastore file [datastore1] 3029c06f-4687-46f7-958b-831e4d0b241b {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1028.479540] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7d266fc6-e2e9-433c-9957-8df9cd0f08b6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.489212] env[63202]: DEBUG oslo_vmware.api [None req-bd8645bc-e2f4-4a6b-a7b2-75b30ea9aa23 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Waiting for the task: (returnval){ [ 1028.489212] env[63202]: value = "task-1385787" [ 1028.489212] env[63202]: _type = "Task" [ 1028.489212] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.499182] env[63202]: DEBUG oslo_vmware.api [None req-bd8645bc-e2f4-4a6b-a7b2-75b30ea9aa23 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385787, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.608769] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385785, 'name': CreateVM_Task, 'duration_secs': 0.446069} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.608985] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1028.609744] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.609912] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.610303] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1028.610594] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-912216d3-5b78-48ba-8e6b-c4163d93baf2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.617736] env[63202]: DEBUG oslo_vmware.api [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 1028.617736] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]5286e66b-9302-6fa4-aaa3-fc3c03ede51c" [ 1028.617736] env[63202]: _type = "Task" [ 1028.617736] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.630829] env[63202]: DEBUG oslo_vmware.api [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]5286e66b-9302-6fa4-aaa3-fc3c03ede51c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.678272] env[63202]: DEBUG oslo_concurrency.lockutils [None req-41f6cd42-ad0d-4d98-b912-a0197bfe098f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.678660] env[63202]: DEBUG oslo_concurrency.lockutils [None req-41f6cd42-ad0d-4d98-b912-a0197bfe098f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.679079] env[63202]: DEBUG nova.objects.instance [None req-41f6cd42-ad0d-4d98-b912-a0197bfe098f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lazy-loading 'resources' on Instance uuid 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1028.869167] env[63202]: DEBUG nova.network.neutron [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Successfully updated port: ea5e41d2-19a6-4899-9bfa-c9b18de354e6 {{(pid=63202) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1028.997365] env[63202]: DEBUG nova.network.neutron [req-026dd47b-4fe1-4f12-b1ff-0949d85df426 req-221f1ad1-2961-48f2-9985-4110437bff90 service nova] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Updated VIF entry in instance network info cache for port 5f6faad3-ee63-4318-abef-e6442d090681. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1028.997836] env[63202]: DEBUG nova.network.neutron [req-026dd47b-4fe1-4f12-b1ff-0949d85df426 req-221f1ad1-2961-48f2-9985-4110437bff90 service nova] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Updating instance_info_cache with network_info: [{"id": "5f6faad3-ee63-4318-abef-e6442d090681", "address": "fa:16:3e:88:4a:c0", "network": {"id": "5a3b4c9b-2ca7-4f8b-8bbb-ea000db91402", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-543677750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f508ec04edc844a19640a8a85f27e5b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f6faad3-ee", "ovs_interfaceid": "5f6faad3-ee63-4318-abef-e6442d090681", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.003394] env[63202]: DEBUG oslo_vmware.api [None req-bd8645bc-e2f4-4a6b-a7b2-75b30ea9aa23 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Task: {'id': task-1385787, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.208255} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.003936] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd8645bc-e2f4-4a6b-a7b2-75b30ea9aa23 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1029.004224] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-bd8645bc-e2f4-4a6b-a7b2-75b30ea9aa23 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Deleted contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1029.004467] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-bd8645bc-e2f4-4a6b-a7b2-75b30ea9aa23 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1029.004729] env[63202]: INFO nova.compute.manager [None req-bd8645bc-e2f4-4a6b-a7b2-75b30ea9aa23 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1029.005044] env[63202]: DEBUG oslo.service.loopingcall [None req-bd8645bc-e2f4-4a6b-a7b2-75b30ea9aa23 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1029.005295] env[63202]: DEBUG nova.compute.manager [-] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1029.005434] env[63202]: DEBUG nova.network.neutron [-] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1029.129468] env[63202]: DEBUG oslo_vmware.api [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]5286e66b-9302-6fa4-aaa3-fc3c03ede51c, 'name': SearchDatastore_Task, 'duration_secs': 0.014118} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.129806] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.130104] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1029.130511] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.130612] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.130753] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1029.131105] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-79a77c5b-0bc0-46dd-a3a6-491e6edef330 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.140906] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1029.141178] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1029.141961] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-176c6f65-93fd-4387-a76b-908f32c07b45 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.148625] env[63202]: DEBUG oslo_vmware.api [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 1029.148625] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]520317f9-4a0d-d516-45cf-31666e54bf80" [ 1029.148625] env[63202]: _type = "Task" [ 1029.148625] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.157952] env[63202]: DEBUG oslo_vmware.api [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]520317f9-4a0d-d516-45cf-31666e54bf80, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.321047] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b110f64-3cc7-41dc-bb75-b13cac0a8950 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.329921] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2741e43c-ac91-4543-896a-f3b5904aea4a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.362810] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f62e1c4-d3bb-4762-a3ae-ec865b0a5096 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.367732] env[63202]: DEBUG nova.compute.manager [req-7600446e-ee0e-4378-b147-5b85872706fb req-7e0a919d-1e76-410e-b68e-299a5327870f service nova] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Received event network-vif-deleted-44f7adb9-2f63-4a9a-a041-0a0ae57c78c8 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1029.367972] env[63202]: INFO nova.compute.manager [req-7600446e-ee0e-4378-b147-5b85872706fb req-7e0a919d-1e76-410e-b68e-299a5327870f service nova] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Neutron deleted interface 44f7adb9-2f63-4a9a-a041-0a0ae57c78c8; detaching it from the instance and deleting it from the info cache [ 1029.368194] env[63202]: DEBUG nova.network.neutron [req-7600446e-ee0e-4378-b147-5b85872706fb req-7e0a919d-1e76-410e-b68e-299a5327870f service nova] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.374635] env[63202]: DEBUG oslo_concurrency.lockutils [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "refresh_cache-c2b6f3a4-7414-4fc4-893b-9b613fc5e381" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.374938] env[63202]: DEBUG oslo_concurrency.lockutils [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquired lock "refresh_cache-c2b6f3a4-7414-4fc4-893b-9b613fc5e381" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.375283] env[63202]: DEBUG nova.network.neutron [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1029.381200] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ec43058-3678-458f-8c92-c8cac82da37c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.397348] env[63202]: DEBUG nova.compute.provider_tree [None req-41f6cd42-ad0d-4d98-b912-a0197bfe098f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1029.504476] env[63202]: DEBUG oslo_concurrency.lockutils [req-026dd47b-4fe1-4f12-b1ff-0949d85df426 req-221f1ad1-2961-48f2-9985-4110437bff90 service nova] Releasing lock "refresh_cache-7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.659234] env[63202]: DEBUG oslo_vmware.api [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]520317f9-4a0d-d516-45cf-31666e54bf80, 'name': SearchDatastore_Task, 'duration_secs': 0.012475} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.660089] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-709523fb-ecd4-4392-96f5-7adc06a91491 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.666606] env[63202]: DEBUG oslo_vmware.api [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 1029.666606] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52df04c1-f96c-9306-0ed6-d1ad29b763fe" [ 1029.666606] env[63202]: _type = "Task" [ 1029.666606] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.674875] env[63202]: DEBUG oslo_vmware.api [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52df04c1-f96c-9306-0ed6-d1ad29b763fe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.756953] env[63202]: DEBUG nova.network.neutron [-] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.872375] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6b7b9677-76d0-4816-9c15-d085547434b1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.885601] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37194bfe-1e68-46a0-b3a7-2844b4e81ff4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.900664] env[63202]: DEBUG nova.scheduler.client.report [None req-41f6cd42-ad0d-4d98-b912-a0197bfe098f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1029.916165] env[63202]: DEBUG nova.compute.manager [req-7600446e-ee0e-4378-b147-5b85872706fb req-7e0a919d-1e76-410e-b68e-299a5327870f service nova] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Detach interface failed, port_id=44f7adb9-2f63-4a9a-a041-0a0ae57c78c8, reason: Instance 3029c06f-4687-46f7-958b-831e4d0b241b could not be found. {{(pid=63202) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1029.927116] env[63202]: DEBUG nova.network.neutron [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1029.948488] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "3a65ddea-2d3d-45a1-a175-c8fad756793d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.948782] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "3a65ddea-2d3d-45a1-a175-c8fad756793d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.077535] env[63202]: DEBUG nova.network.neutron [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Updating instance_info_cache with network_info: [{"id": "ea5e41d2-19a6-4899-9bfa-c9b18de354e6", "address": "fa:16:3e:28:de:2f", "network": {"id": "18390479-16ce-4012-9ba6-abf19b5616d1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1376026898-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22352c25bca8416a948014391a5389ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "085fb0ff-9285-4f1d-a008-a14da4844357", "external-id": "nsx-vlan-transportzone-729", "segmentation_id": 729, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea5e41d2-19", "ovs_interfaceid": "ea5e41d2-19a6-4899-9bfa-c9b18de354e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.179045] env[63202]: DEBUG oslo_vmware.api [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52df04c1-f96c-9306-0ed6-d1ad29b763fe, 'name': SearchDatastore_Task, 'duration_secs': 0.011944} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.179224] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.179474] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b/7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1030.179745] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b4848443-20ba-448c-a509-2af280a120df {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.187804] env[63202]: DEBUG oslo_vmware.api [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 1030.187804] env[63202]: value = "task-1385788" [ 1030.187804] env[63202]: _type = "Task" [ 1030.187804] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.196078] env[63202]: DEBUG oslo_vmware.api [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385788, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.258755] env[63202]: INFO nova.compute.manager [-] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Took 1.25 seconds to deallocate network for instance. [ 1030.344508] env[63202]: DEBUG nova.compute.manager [req-79e47e5b-454f-4785-9065-06b31a2a09c0 req-7a36c0e7-9371-45cd-97b4-eb119a935f54 service nova] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Received event network-vif-plugged-ea5e41d2-19a6-4899-9bfa-c9b18de354e6 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1030.344508] env[63202]: DEBUG oslo_concurrency.lockutils [req-79e47e5b-454f-4785-9065-06b31a2a09c0 req-7a36c0e7-9371-45cd-97b4-eb119a935f54 service nova] Acquiring lock "c2b6f3a4-7414-4fc4-893b-9b613fc5e381-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.344508] env[63202]: DEBUG oslo_concurrency.lockutils [req-79e47e5b-454f-4785-9065-06b31a2a09c0 req-7a36c0e7-9371-45cd-97b4-eb119a935f54 service nova] Lock "c2b6f3a4-7414-4fc4-893b-9b613fc5e381-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.344508] env[63202]: DEBUG oslo_concurrency.lockutils [req-79e47e5b-454f-4785-9065-06b31a2a09c0 req-7a36c0e7-9371-45cd-97b4-eb119a935f54 service nova] Lock "c2b6f3a4-7414-4fc4-893b-9b613fc5e381-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.344969] env[63202]: DEBUG nova.compute.manager [req-79e47e5b-454f-4785-9065-06b31a2a09c0 req-7a36c0e7-9371-45cd-97b4-eb119a935f54 service nova] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] No waiting events found dispatching network-vif-plugged-ea5e41d2-19a6-4899-9bfa-c9b18de354e6 {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1030.344969] env[63202]: WARNING nova.compute.manager [req-79e47e5b-454f-4785-9065-06b31a2a09c0 req-7a36c0e7-9371-45cd-97b4-eb119a935f54 service nova] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Received unexpected event network-vif-plugged-ea5e41d2-19a6-4899-9bfa-c9b18de354e6 for instance with vm_state building and task_state spawning. [ 1030.344969] env[63202]: DEBUG nova.compute.manager [req-79e47e5b-454f-4785-9065-06b31a2a09c0 req-7a36c0e7-9371-45cd-97b4-eb119a935f54 service nova] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Received event network-changed-ea5e41d2-19a6-4899-9bfa-c9b18de354e6 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1030.345566] env[63202]: DEBUG nova.compute.manager [req-79e47e5b-454f-4785-9065-06b31a2a09c0 req-7a36c0e7-9371-45cd-97b4-eb119a935f54 service nova] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Refreshing instance network info cache due to event network-changed-ea5e41d2-19a6-4899-9bfa-c9b18de354e6. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1030.345566] env[63202]: DEBUG oslo_concurrency.lockutils [req-79e47e5b-454f-4785-9065-06b31a2a09c0 req-7a36c0e7-9371-45cd-97b4-eb119a935f54 service nova] Acquiring lock "refresh_cache-c2b6f3a4-7414-4fc4-893b-9b613fc5e381" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.408035] env[63202]: DEBUG oslo_concurrency.lockutils [None req-41f6cd42-ad0d-4d98-b912-a0197bfe098f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.729s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.432760] env[63202]: INFO nova.scheduler.client.report [None req-41f6cd42-ad0d-4d98-b912-a0197bfe098f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Deleted allocations for instance 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb [ 1030.451783] env[63202]: DEBUG nova.compute.manager [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1030.581170] env[63202]: DEBUG oslo_concurrency.lockutils [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Releasing lock "refresh_cache-c2b6f3a4-7414-4fc4-893b-9b613fc5e381" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.581613] env[63202]: DEBUG nova.compute.manager [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Instance network_info: |[{"id": "ea5e41d2-19a6-4899-9bfa-c9b18de354e6", "address": "fa:16:3e:28:de:2f", "network": {"id": "18390479-16ce-4012-9ba6-abf19b5616d1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1376026898-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22352c25bca8416a948014391a5389ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "085fb0ff-9285-4f1d-a008-a14da4844357", "external-id": "nsx-vlan-transportzone-729", "segmentation_id": 729, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea5e41d2-19", "ovs_interfaceid": "ea5e41d2-19a6-4899-9bfa-c9b18de354e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1030.582019] env[63202]: DEBUG oslo_concurrency.lockutils [req-79e47e5b-454f-4785-9065-06b31a2a09c0 req-7a36c0e7-9371-45cd-97b4-eb119a935f54 service nova] Acquired lock "refresh_cache-c2b6f3a4-7414-4fc4-893b-9b613fc5e381" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.582298] env[63202]: DEBUG nova.network.neutron [req-79e47e5b-454f-4785-9065-06b31a2a09c0 req-7a36c0e7-9371-45cd-97b4-eb119a935f54 service nova] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Refreshing network info cache for port ea5e41d2-19a6-4899-9bfa-c9b18de354e6 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1030.584315] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:28:de:2f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '085fb0ff-9285-4f1d-a008-a14da4844357', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ea5e41d2-19a6-4899-9bfa-c9b18de354e6', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1030.592265] env[63202]: DEBUG oslo.service.loopingcall [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1030.593316] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1030.594052] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-30f1733e-58fc-4646-9d78-6bf7a6cf6d48 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.617598] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1030.617598] env[63202]: value = "task-1385789" [ 1030.617598] env[63202]: _type = "Task" [ 1030.617598] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.702335] env[63202]: DEBUG oslo_vmware.api [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385788, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.766635] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bd8645bc-e2f4-4a6b-a7b2-75b30ea9aa23 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.767132] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bd8645bc-e2f4-4a6b-a7b2-75b30ea9aa23 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.767419] env[63202]: DEBUG nova.objects.instance [None req-bd8645bc-e2f4-4a6b-a7b2-75b30ea9aa23 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Lazy-loading 'resources' on Instance uuid 3029c06f-4687-46f7-958b-831e4d0b241b {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1030.943313] env[63202]: DEBUG oslo_concurrency.lockutils [None req-41f6cd42-ad0d-4d98-b912-a0197bfe098f tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lock "0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.349s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.974000] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.128466] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385789, 'name': CreateVM_Task, 'duration_secs': 0.438394} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.128672] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1031.129393] env[63202]: DEBUG oslo_concurrency.lockutils [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1031.129567] env[63202]: DEBUG oslo_concurrency.lockutils [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.129909] env[63202]: DEBUG oslo_concurrency.lockutils [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1031.130332] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-259fc580-2049-4f84-9944-60252c793bb7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.135990] env[63202]: DEBUG oslo_vmware.api [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for the task: (returnval){ [ 1031.135990] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]5235c48f-b02a-2877-fb1d-8786b0e63c3f" [ 1031.135990] env[63202]: _type = "Task" [ 1031.135990] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.147898] env[63202]: DEBUG oslo_vmware.api [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]5235c48f-b02a-2877-fb1d-8786b0e63c3f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.200809] env[63202]: DEBUG oslo_vmware.api [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385788, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.385247] env[63202]: DEBUG nova.network.neutron [req-79e47e5b-454f-4785-9065-06b31a2a09c0 req-7a36c0e7-9371-45cd-97b4-eb119a935f54 service nova] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Updated VIF entry in instance network info cache for port ea5e41d2-19a6-4899-9bfa-c9b18de354e6. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1031.385622] env[63202]: DEBUG nova.network.neutron [req-79e47e5b-454f-4785-9065-06b31a2a09c0 req-7a36c0e7-9371-45cd-97b4-eb119a935f54 service nova] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Updating instance_info_cache with network_info: [{"id": "ea5e41d2-19a6-4899-9bfa-c9b18de354e6", "address": "fa:16:3e:28:de:2f", "network": {"id": "18390479-16ce-4012-9ba6-abf19b5616d1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1376026898-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22352c25bca8416a948014391a5389ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "085fb0ff-9285-4f1d-a008-a14da4844357", "external-id": "nsx-vlan-transportzone-729", "segmentation_id": 729, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea5e41d2-19", "ovs_interfaceid": "ea5e41d2-19a6-4899-9bfa-c9b18de354e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1031.408446] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90609ade-1868-47c8-8b87-b77c8c7b9144 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.417524] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd7bbdbc-2015-42f8-af77-9812deff2ecf {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.451610] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14fdd9da-a19c-4197-ab1a-dd3a802c1a25 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.461758] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c3375d5-7bf5-4596-8ff0-b4a3fff309a1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.475138] env[63202]: DEBUG nova.compute.provider_tree [None req-bd8645bc-e2f4-4a6b-a7b2-75b30ea9aa23 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1031.647678] env[63202]: DEBUG oslo_vmware.api [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]5235c48f-b02a-2877-fb1d-8786b0e63c3f, 'name': SearchDatastore_Task, 'duration_secs': 0.014286} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.648066] env[63202]: DEBUG oslo_concurrency.lockutils [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.648239] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1031.648481] env[63202]: DEBUG oslo_concurrency.lockutils [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1031.648624] env[63202]: DEBUG oslo_concurrency.lockutils [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.648801] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1031.649132] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-71c26d57-d604-425c-8ebb-e665068104fe {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.664417] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1031.664725] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1031.665984] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-be2ca9d9-71fc-4121-8b32-20402f2c9be1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.674583] env[63202]: DEBUG oslo_vmware.api [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for the task: (returnval){ [ 1031.674583] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52c2a774-943e-a8d7-b584-6875ae9dfdf6" [ 1031.674583] env[63202]: _type = "Task" [ 1031.674583] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.684069] env[63202]: DEBUG oslo_vmware.api [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52c2a774-943e-a8d7-b584-6875ae9dfdf6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.695559] env[63202]: DEBUG oslo_vmware.rw_handles [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b09717-c9a4-f5dc-9b54-1c5cd758adb9/disk-0.vmdk. {{(pid=63202) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1031.696386] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-559c19d1-1e5b-4bfc-ae31-adf8c78913c9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.703125] env[63202]: DEBUG oslo_vmware.rw_handles [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b09717-c9a4-f5dc-9b54-1c5cd758adb9/disk-0.vmdk is in state: ready. {{(pid=63202) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1031.703314] env[63202]: ERROR oslo_vmware.rw_handles [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b09717-c9a4-f5dc-9b54-1c5cd758adb9/disk-0.vmdk due to incomplete transfer. [ 1031.706478] env[63202]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-2d544aec-14f6-4c69-a6b2-ac0c34d5b006 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.708108] env[63202]: DEBUG oslo_vmware.api [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385788, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.506071} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.708359] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b/7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1031.708581] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1031.709148] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4ada90c8-8b36-42f2-952d-74fe3d21c3b2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.714178] env[63202]: DEBUG oslo_vmware.rw_handles [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b09717-c9a4-f5dc-9b54-1c5cd758adb9/disk-0.vmdk. {{(pid=63202) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1031.714365] env[63202]: DEBUG nova.virt.vmwareapi.images [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Uploaded image 3db9d5e2-204c-47e7-ae45-3ed10283296d to the Glance image server {{(pid=63202) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1031.716805] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Destroying the VM {{(pid=63202) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1031.718043] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-de5d6db7-d450-43b1-b5e8-526697791018 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.719500] env[63202]: DEBUG oslo_vmware.api [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 1031.719500] env[63202]: value = "task-1385790" [ 1031.719500] env[63202]: _type = "Task" [ 1031.719500] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.730341] env[63202]: DEBUG oslo_vmware.api [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1031.730341] env[63202]: value = "task-1385791" [ 1031.730341] env[63202]: _type = "Task" [ 1031.730341] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.736483] env[63202]: DEBUG oslo_vmware.api [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385790, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.742150] env[63202]: DEBUG oslo_vmware.api [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385791, 'name': Destroy_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.892607] env[63202]: DEBUG oslo_concurrency.lockutils [req-79e47e5b-454f-4785-9065-06b31a2a09c0 req-7a36c0e7-9371-45cd-97b4-eb119a935f54 service nova] Releasing lock "refresh_cache-c2b6f3a4-7414-4fc4-893b-9b613fc5e381" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.979203] env[63202]: DEBUG nova.scheduler.client.report [None req-bd8645bc-e2f4-4a6b-a7b2-75b30ea9aa23 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1032.186018] env[63202]: DEBUG oslo_vmware.api [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52c2a774-943e-a8d7-b584-6875ae9dfdf6, 'name': SearchDatastore_Task, 'duration_secs': 0.017456} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.186924] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5e24b577-1ea0-4028-b2f8-1e6dc255a21e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.192944] env[63202]: DEBUG oslo_vmware.api [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for the task: (returnval){ [ 1032.192944] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52c3b7b7-e6c1-1de3-a709-483f9c659c2d" [ 1032.192944] env[63202]: _type = "Task" [ 1032.192944] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.201111] env[63202]: DEBUG oslo_vmware.api [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52c3b7b7-e6c1-1de3-a709-483f9c659c2d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.229361] env[63202]: DEBUG oslo_vmware.api [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385790, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079675} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.230684] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1032.231301] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b1ba85b-fa0e-4eb6-8de6-92eeb801ed67 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.245325] env[63202]: DEBUG oslo_vmware.api [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385791, 'name': Destroy_Task, 'duration_secs': 0.313585} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.255996] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Destroyed the VM [ 1032.256332] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Deleting Snapshot of the VM instance {{(pid=63202) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1032.267265] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Reconfiguring VM instance instance-0000005f to attach disk [datastore1] 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b/7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1032.267794] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-e9083827-321b-4c76-9239-02d1dbbdd695 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.271417] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-47f68a7a-71f8-4205-bd84-26fc14d97e6c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.294244] env[63202]: DEBUG oslo_vmware.api [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 1032.294244] env[63202]: value = "task-1385793" [ 1032.294244] env[63202]: _type = "Task" [ 1032.294244] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.295701] env[63202]: DEBUG oslo_vmware.api [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1032.295701] env[63202]: value = "task-1385792" [ 1032.295701] env[63202]: _type = "Task" [ 1032.295701] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.314276] env[63202]: DEBUG oslo_vmware.api [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385793, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.316881] env[63202]: DEBUG oslo_vmware.api [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385792, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.484839] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bd8645bc-e2f4-4a6b-a7b2-75b30ea9aa23 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.718s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.487355] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.513s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.488958] env[63202]: INFO nova.compute.claims [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1032.501386] env[63202]: INFO nova.scheduler.client.report [None req-bd8645bc-e2f4-4a6b-a7b2-75b30ea9aa23 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Deleted allocations for instance 3029c06f-4687-46f7-958b-831e4d0b241b [ 1032.703408] env[63202]: DEBUG oslo_vmware.api [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52c3b7b7-e6c1-1de3-a709-483f9c659c2d, 'name': SearchDatastore_Task, 'duration_secs': 0.010194} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.703748] env[63202]: DEBUG oslo_concurrency.lockutils [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1032.704035] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] c2b6f3a4-7414-4fc4-893b-9b613fc5e381/c2b6f3a4-7414-4fc4-893b-9b613fc5e381.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1032.704337] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-73653a1b-de37-452f-9175-a6eb4d4a9e13 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.713049] env[63202]: DEBUG oslo_vmware.api [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for the task: (returnval){ [ 1032.713049] env[63202]: value = "task-1385794" [ 1032.713049] env[63202]: _type = "Task" [ 1032.713049] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.721617] env[63202]: DEBUG oslo_vmware.api [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385794, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.808268] env[63202]: DEBUG oslo_vmware.api [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385793, 'name': ReconfigVM_Task, 'duration_secs': 0.312133} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.811739] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Reconfigured VM instance instance-0000005f to attach disk [datastore1] 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b/7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1032.812483] env[63202]: DEBUG oslo_vmware.api [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385792, 'name': RemoveSnapshot_Task, 'duration_secs': 0.373069} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.812724] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-217a10b3-2eca-4ef4-9d9c-578c7c20d761 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.814529] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Deleted Snapshot of the VM instance {{(pid=63202) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1032.814724] env[63202]: DEBUG nova.compute.manager [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1032.815518] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36413a53-6a8a-4d5c-9c36-5ecfea816852 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.826401] env[63202]: DEBUG oslo_vmware.api [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 1032.826401] env[63202]: value = "task-1385795" [ 1032.826401] env[63202]: _type = "Task" [ 1032.826401] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.839767] env[63202]: DEBUG oslo_vmware.api [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385795, 'name': Rename_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.846456] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Acquiring lock "e1e2e8a4-80aa-45eb-a90e-d20cba2943ee" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.846703] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lock "e1e2e8a4-80aa-45eb-a90e-d20cba2943ee" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.008265] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bd8645bc-e2f4-4a6b-a7b2-75b30ea9aa23 tempest-ImagesOneServerNegativeTestJSON-1139626155 tempest-ImagesOneServerNegativeTestJSON-1139626155-project-member] Lock "3029c06f-4687-46f7-958b-831e4d0b241b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.149s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.224228] env[63202]: DEBUG oslo_vmware.api [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385794, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.437548} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.224421] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] c2b6f3a4-7414-4fc4-893b-9b613fc5e381/c2b6f3a4-7414-4fc4-893b-9b613fc5e381.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1033.224653] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1033.224943] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7d8ae3fd-46b1-4e50-b9d7-229ab37c6ae7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.231700] env[63202]: DEBUG oslo_vmware.api [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for the task: (returnval){ [ 1033.231700] env[63202]: value = "task-1385796" [ 1033.231700] env[63202]: _type = "Task" [ 1033.231700] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.240299] env[63202]: DEBUG oslo_vmware.api [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385796, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.334400] env[63202]: INFO nova.compute.manager [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Shelve offloading [ 1033.335986] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1033.340029] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ac611113-1591-404c-a91d-b8134cc64694 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.341875] env[63202]: DEBUG oslo_vmware.api [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385795, 'name': Rename_Task, 'duration_secs': 0.399395} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.342241] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1033.343677] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-95a307d8-e597-47a3-85cf-7d0d1b7335a6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.345611] env[63202]: DEBUG oslo_vmware.api [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1033.345611] env[63202]: value = "task-1385797" [ 1033.345611] env[63202]: _type = "Task" [ 1033.345611] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.350730] env[63202]: DEBUG nova.compute.manager [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1033.358189] env[63202]: DEBUG oslo_vmware.api [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 1033.358189] env[63202]: value = "task-1385798" [ 1033.358189] env[63202]: _type = "Task" [ 1033.358189] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.363860] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] VM already powered off {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1033.363984] env[63202]: DEBUG nova.compute.manager [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1033.364880] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc91d1ef-3ccd-428d-8652-5e93b91352ad {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.372670] env[63202]: DEBUG oslo_vmware.api [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385798, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.378882] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquiring lock "refresh_cache-8437f856-a707-49c5-b8eb-5a22cdb990f8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.378882] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquired lock "refresh_cache-8437f856-a707-49c5-b8eb-5a22cdb990f8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.378882] env[63202]: DEBUG nova.network.neutron [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1033.597569] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de32a32c-9ece-4118-aef2-e7eafb5dddaa {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.606913] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b87857cf-e4a7-4dc4-89d4-8bcefe6baf92 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.641421] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be57311a-f45d-4ff3-a184-f3700d9f3220 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.649475] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b1a138e-6a46-42a1-b619-a536a3672d26 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.664807] env[63202]: DEBUG nova.compute.provider_tree [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1033.742688] env[63202]: DEBUG oslo_vmware.api [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385796, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068897} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.743041] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1033.743846] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5336b483-655c-413a-84d0-a6173e20f3d1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.765876] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Reconfiguring VM instance instance-00000060 to attach disk [datastore1] c2b6f3a4-7414-4fc4-893b-9b613fc5e381/c2b6f3a4-7414-4fc4-893b-9b613fc5e381.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1033.766215] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7d858fb1-7511-4dde-a665-25c7f32109d8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.785461] env[63202]: DEBUG oslo_vmware.api [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for the task: (returnval){ [ 1033.785461] env[63202]: value = "task-1385799" [ 1033.785461] env[63202]: _type = "Task" [ 1033.785461] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.793511] env[63202]: DEBUG oslo_vmware.api [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385799, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.868986] env[63202]: DEBUG oslo_vmware.api [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385798, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.873805] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.157737] env[63202]: DEBUG nova.network.neutron [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Updating instance_info_cache with network_info: [{"id": "07cc827a-876a-4432-a5fc-9ba8920dc5f9", "address": "fa:16:3e:71:e6:9f", "network": {"id": "06ab5813-9ad9-4021-9bdb-f2f02af8d73f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1714653503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.175", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b2de35030a9484094e964ffc30a822d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3093647a-bab7-4562-ada0-428725e8c0fc", "external-id": "nsx-vlan-transportzone-660", "segmentation_id": 660, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap07cc827a-87", "ovs_interfaceid": "07cc827a-876a-4432-a5fc-9ba8920dc5f9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.168272] env[63202]: DEBUG nova.scheduler.client.report [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1034.297398] env[63202]: DEBUG oslo_vmware.api [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385799, 'name': ReconfigVM_Task, 'duration_secs': 0.315907} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.297805] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Reconfigured VM instance instance-00000060 to attach disk [datastore1] c2b6f3a4-7414-4fc4-893b-9b613fc5e381/c2b6f3a4-7414-4fc4-893b-9b613fc5e381.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1034.298562] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-13deb2ea-aa6e-474e-8d14-6893974df9b9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.308216] env[63202]: DEBUG oslo_vmware.api [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for the task: (returnval){ [ 1034.308216] env[63202]: value = "task-1385800" [ 1034.308216] env[63202]: _type = "Task" [ 1034.308216] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.318850] env[63202]: DEBUG oslo_vmware.api [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385800, 'name': Rename_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.367829] env[63202]: DEBUG oslo_vmware.api [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385798, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.660876] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Releasing lock "refresh_cache-8437f856-a707-49c5-b8eb-5a22cdb990f8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.672890] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.185s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.673486] env[63202]: DEBUG nova.compute.manager [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1034.676522] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.803s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.678290] env[63202]: INFO nova.compute.claims [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1034.818630] env[63202]: DEBUG oslo_vmware.api [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385800, 'name': Rename_Task, 'duration_secs': 0.170673} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.818918] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1034.819433] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-054995f0-e74e-4f45-9ebc-f224c31587ad {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.828411] env[63202]: DEBUG oslo_vmware.api [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for the task: (returnval){ [ 1034.828411] env[63202]: value = "task-1385801" [ 1034.828411] env[63202]: _type = "Task" [ 1034.828411] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.837215] env[63202]: DEBUG oslo_vmware.api [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385801, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.877859] env[63202]: DEBUG oslo_vmware.api [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385798, 'name': PowerOnVM_Task, 'duration_secs': 1.437869} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.879467] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1034.879710] env[63202]: INFO nova.compute.manager [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Took 9.43 seconds to spawn the instance on the hypervisor. [ 1034.880065] env[63202]: DEBUG nova.compute.manager [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1034.881309] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c561479f-6c1f-4b0b-a923-5e8dfd9f929a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.994218] env[63202]: DEBUG nova.compute.manager [req-ee2f444f-5c5f-4ba3-822f-e0ba2d152a8d req-460ac11c-b367-40cc-9251-c05421632ddb service nova] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Received event network-vif-unplugged-07cc827a-876a-4432-a5fc-9ba8920dc5f9 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1034.994471] env[63202]: DEBUG oslo_concurrency.lockutils [req-ee2f444f-5c5f-4ba3-822f-e0ba2d152a8d req-460ac11c-b367-40cc-9251-c05421632ddb service nova] Acquiring lock "8437f856-a707-49c5-b8eb-5a22cdb990f8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.994788] env[63202]: DEBUG oslo_concurrency.lockutils [req-ee2f444f-5c5f-4ba3-822f-e0ba2d152a8d req-460ac11c-b367-40cc-9251-c05421632ddb service nova] Lock "8437f856-a707-49c5-b8eb-5a22cdb990f8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.995045] env[63202]: DEBUG oslo_concurrency.lockutils [req-ee2f444f-5c5f-4ba3-822f-e0ba2d152a8d req-460ac11c-b367-40cc-9251-c05421632ddb service nova] Lock "8437f856-a707-49c5-b8eb-5a22cdb990f8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.995353] env[63202]: DEBUG nova.compute.manager [req-ee2f444f-5c5f-4ba3-822f-e0ba2d152a8d req-460ac11c-b367-40cc-9251-c05421632ddb service nova] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] No waiting events found dispatching network-vif-unplugged-07cc827a-876a-4432-a5fc-9ba8920dc5f9 {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1034.995566] env[63202]: WARNING nova.compute.manager [req-ee2f444f-5c5f-4ba3-822f-e0ba2d152a8d req-460ac11c-b367-40cc-9251-c05421632ddb service nova] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Received unexpected event network-vif-unplugged-07cc827a-876a-4432-a5fc-9ba8920dc5f9 for instance with vm_state shelved and task_state shelving_offloading. [ 1035.088443] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1035.089511] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dceec2f-efb3-49b4-b06a-5a011b44fe0d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.098582] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1035.098785] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f08dd85c-4549-47e2-9483-d169f1178618 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.168280] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1035.168724] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Deleting contents of the VM from datastore datastore2 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1035.169031] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Deleting the datastore file [datastore2] 8437f856-a707-49c5-b8eb-5a22cdb990f8 {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1035.169362] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c4dc7527-9c32-41f7-a558-e6c248a4851c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.178712] env[63202]: DEBUG nova.compute.utils [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1035.180881] env[63202]: DEBUG oslo_vmware.api [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1035.180881] env[63202]: value = "task-1385803" [ 1035.180881] env[63202]: _type = "Task" [ 1035.180881] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.181238] env[63202]: DEBUG nova.compute.manager [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1035.181483] env[63202]: DEBUG nova.network.neutron [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1035.196843] env[63202]: DEBUG oslo_vmware.api [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385803, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.230223] env[63202]: DEBUG nova.policy [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2bd91973e1bb4ce7be8bce9bfe34a941', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e972faeaedc6468aab7e7cfee88a477b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 1035.339415] env[63202]: DEBUG oslo_vmware.api [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385801, 'name': PowerOnVM_Task, 'duration_secs': 0.476167} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.340807] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1035.340917] env[63202]: INFO nova.compute.manager [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Took 7.15 seconds to spawn the instance on the hypervisor. [ 1035.341107] env[63202]: DEBUG nova.compute.manager [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1035.343465] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f964181-ead6-41f2-80f8-d62c099d8388 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.400249] env[63202]: INFO nova.compute.manager [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Took 14.22 seconds to build instance. [ 1035.584245] env[63202]: DEBUG nova.network.neutron [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Successfully created port: 3bfc4703-af3e-41fe-a35a-a5a8fadf1aef {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1035.685218] env[63202]: DEBUG nova.compute.manager [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1035.701470] env[63202]: DEBUG oslo_vmware.api [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385803, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.247633} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.701732] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1035.701912] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Deleted contents of the VM from datastore datastore2 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1035.702100] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1035.724174] env[63202]: INFO nova.scheduler.client.report [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Deleted allocations for instance 8437f856-a707-49c5-b8eb-5a22cdb990f8 [ 1035.831566] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee0ffef7-e15d-4b2f-9701-f1a01379586f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.840618] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f91b6fb-dbf5-4714-9542-1dab0c57f725 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.847438] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2191f03e-0414-4aaf-bd3b-8677a00c381d tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.882496] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a61a5327-1b2d-4876-9eb7-79c52c038f1c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.886942] env[63202]: INFO nova.compute.manager [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Took 12.01 seconds to build instance. [ 1035.895312] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbfd6ac0-c008-4d64-b656-cd78fd1cd506 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.902582] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2e2faec7-b2c3-4cb8-842d-aaf07f7784d7 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.733s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.910821] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2191f03e-0414-4aaf-bd3b-8677a00c381d tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.063s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.911054] env[63202]: DEBUG nova.compute.manager [None req-2191f03e-0414-4aaf-bd3b-8677a00c381d tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1035.911544] env[63202]: DEBUG nova.compute.provider_tree [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1035.913967] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-927e8a61-e10c-49c2-bce1-ecb30b0db17f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.921274] env[63202]: DEBUG nova.compute.manager [None req-2191f03e-0414-4aaf-bd3b-8677a00c381d tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63202) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1035.921862] env[63202]: DEBUG nova.objects.instance [None req-2191f03e-0414-4aaf-bd3b-8677a00c381d tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lazy-loading 'flavor' on Instance uuid 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1036.231904] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.389313] env[63202]: DEBUG oslo_concurrency.lockutils [None req-da56ebba-ffb2-4b69-ad66-48ac9346af00 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "c2b6f3a4-7414-4fc4-893b-9b613fc5e381" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.519s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.418940] env[63202]: DEBUG nova.scheduler.client.report [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1036.432025] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-2191f03e-0414-4aaf-bd3b-8677a00c381d tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1036.432025] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9200d042-3f7d-4155-a8cf-3ce16fe3cb84 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.440383] env[63202]: DEBUG oslo_vmware.api [None req-2191f03e-0414-4aaf-bd3b-8677a00c381d tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 1036.440383] env[63202]: value = "task-1385804" [ 1036.440383] env[63202]: _type = "Task" [ 1036.440383] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.452931] env[63202]: DEBUG oslo_vmware.api [None req-2191f03e-0414-4aaf-bd3b-8677a00c381d tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385804, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.695283] env[63202]: DEBUG nova.compute.manager [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1036.725614] env[63202]: DEBUG nova.virt.hardware [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1036.726128] env[63202]: DEBUG nova.virt.hardware [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1036.726448] env[63202]: DEBUG nova.virt.hardware [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1036.726755] env[63202]: DEBUG nova.virt.hardware [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1036.727130] env[63202]: DEBUG nova.virt.hardware [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1036.727426] env[63202]: DEBUG nova.virt.hardware [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1036.727769] env[63202]: DEBUG nova.virt.hardware [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1036.728074] env[63202]: DEBUG nova.virt.hardware [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1036.728369] env[63202]: DEBUG nova.virt.hardware [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1036.728652] env[63202]: DEBUG nova.virt.hardware [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1036.728975] env[63202]: DEBUG nova.virt.hardware [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1036.730544] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05672b7b-b2a3-4e37-8dd3-25d4eb257c81 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.739740] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-660cbc8a-d189-4c23-95d9-61df6771c841 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.924490] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.248s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.925245] env[63202]: DEBUG nova.compute.manager [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1036.927801] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.696s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.928090] env[63202]: DEBUG nova.objects.instance [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lazy-loading 'resources' on Instance uuid 8437f856-a707-49c5-b8eb-5a22cdb990f8 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1036.952080] env[63202]: DEBUG oslo_vmware.api [None req-2191f03e-0414-4aaf-bd3b-8677a00c381d tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385804, 'name': PowerOffVM_Task, 'duration_secs': 0.261754} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.952080] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-2191f03e-0414-4aaf-bd3b-8677a00c381d tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1036.952080] env[63202]: DEBUG nova.compute.manager [None req-2191f03e-0414-4aaf-bd3b-8677a00c381d tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1036.953056] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2c294d0-14b5-49cc-bfcd-c09ab94d6867 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.052831] env[63202]: DEBUG nova.compute.manager [req-393160ad-2ec6-486f-8bf7-771a816076ae req-39164185-9ec9-44fb-a722-e40bb9f9fa35 service nova] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Received event network-changed-07cc827a-876a-4432-a5fc-9ba8920dc5f9 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1037.052831] env[63202]: DEBUG nova.compute.manager [req-393160ad-2ec6-486f-8bf7-771a816076ae req-39164185-9ec9-44fb-a722-e40bb9f9fa35 service nova] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Refreshing instance network info cache due to event network-changed-07cc827a-876a-4432-a5fc-9ba8920dc5f9. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1037.052831] env[63202]: DEBUG oslo_concurrency.lockutils [req-393160ad-2ec6-486f-8bf7-771a816076ae req-39164185-9ec9-44fb-a722-e40bb9f9fa35 service nova] Acquiring lock "refresh_cache-8437f856-a707-49c5-b8eb-5a22cdb990f8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.052831] env[63202]: DEBUG oslo_concurrency.lockutils [req-393160ad-2ec6-486f-8bf7-771a816076ae req-39164185-9ec9-44fb-a722-e40bb9f9fa35 service nova] Acquired lock "refresh_cache-8437f856-a707-49c5-b8eb-5a22cdb990f8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.052831] env[63202]: DEBUG nova.network.neutron [req-393160ad-2ec6-486f-8bf7-771a816076ae req-39164185-9ec9-44fb-a722-e40bb9f9fa35 service nova] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Refreshing network info cache for port 07cc827a-876a-4432-a5fc-9ba8920dc5f9 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1037.063368] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7d0ec686-adfb-458e-8a0e-e648f13a89a8 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "c2b6f3a4-7414-4fc4-893b-9b613fc5e381" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.063368] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7d0ec686-adfb-458e-8a0e-e648f13a89a8 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "c2b6f3a4-7414-4fc4-893b-9b613fc5e381" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.063368] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7d0ec686-adfb-458e-8a0e-e648f13a89a8 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "c2b6f3a4-7414-4fc4-893b-9b613fc5e381-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.063368] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7d0ec686-adfb-458e-8a0e-e648f13a89a8 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "c2b6f3a4-7414-4fc4-893b-9b613fc5e381-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.063627] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7d0ec686-adfb-458e-8a0e-e648f13a89a8 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "c2b6f3a4-7414-4fc4-893b-9b613fc5e381-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.064337] env[63202]: INFO nova.compute.manager [None req-7d0ec686-adfb-458e-8a0e-e648f13a89a8 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Terminating instance [ 1037.066286] env[63202]: DEBUG nova.compute.manager [None req-7d0ec686-adfb-458e-8a0e-e648f13a89a8 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1037.066844] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-7d0ec686-adfb-458e-8a0e-e648f13a89a8 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1037.068182] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5314cebc-6744-4b76-8fea-bd72ef937be7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.079604] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d0ec686-adfb-458e-8a0e-e648f13a89a8 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1037.080096] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fe8e6716-d8f9-4c53-a643-83f8d3961fbb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.087981] env[63202]: DEBUG oslo_vmware.api [None req-7d0ec686-adfb-458e-8a0e-e648f13a89a8 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for the task: (returnval){ [ 1037.087981] env[63202]: value = "task-1385805" [ 1037.087981] env[63202]: _type = "Task" [ 1037.087981] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.098543] env[63202]: DEBUG oslo_vmware.api [None req-7d0ec686-adfb-458e-8a0e-e648f13a89a8 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385805, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.431231] env[63202]: DEBUG nova.compute.utils [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1037.433122] env[63202]: DEBUG nova.objects.instance [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lazy-loading 'numa_topology' on Instance uuid 8437f856-a707-49c5-b8eb-5a22cdb990f8 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1037.436788] env[63202]: DEBUG nova.compute.manager [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1037.437132] env[63202]: DEBUG nova.network.neutron [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1037.469018] env[63202]: DEBUG nova.network.neutron [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Successfully updated port: 3bfc4703-af3e-41fe-a35a-a5a8fadf1aef {{(pid=63202) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1037.469574] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2191f03e-0414-4aaf-bd3b-8677a00c381d tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.559s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.598702] env[63202]: DEBUG oslo_vmware.api [None req-7d0ec686-adfb-458e-8a0e-e648f13a89a8 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385805, 'name': PowerOffVM_Task, 'duration_secs': 0.17706} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.598979] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d0ec686-adfb-458e-8a0e-e648f13a89a8 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1037.599164] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-7d0ec686-adfb-458e-8a0e-e648f13a89a8 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1037.599423] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-78714da5-5b5b-426a-8381-30bae59eedbf {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.694949] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-7d0ec686-adfb-458e-8a0e-e648f13a89a8 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1037.694949] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-7d0ec686-adfb-458e-8a0e-e648f13a89a8 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Deleting contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1037.695806] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-7d0ec686-adfb-458e-8a0e-e648f13a89a8 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Deleting the datastore file [datastore1] c2b6f3a4-7414-4fc4-893b-9b613fc5e381 {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1037.697435] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c005fe3e-9499-42ef-86bc-2b68628ca4ef {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.707167] env[63202]: DEBUG oslo_vmware.api [None req-7d0ec686-adfb-458e-8a0e-e648f13a89a8 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for the task: (returnval){ [ 1037.707167] env[63202]: value = "task-1385807" [ 1037.707167] env[63202]: _type = "Task" [ 1037.707167] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.720700] env[63202]: DEBUG oslo_vmware.api [None req-7d0ec686-adfb-458e-8a0e-e648f13a89a8 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385807, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.731097] env[63202]: DEBUG nova.policy [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7e11242d52d14577847d17d3a6441202', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a9e939cdf4a94a8aa0107a5761771c2c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 1037.897273] env[63202]: DEBUG nova.network.neutron [req-393160ad-2ec6-486f-8bf7-771a816076ae req-39164185-9ec9-44fb-a722-e40bb9f9fa35 service nova] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Updated VIF entry in instance network info cache for port 07cc827a-876a-4432-a5fc-9ba8920dc5f9. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1037.897662] env[63202]: DEBUG nova.network.neutron [req-393160ad-2ec6-486f-8bf7-771a816076ae req-39164185-9ec9-44fb-a722-e40bb9f9fa35 service nova] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Updating instance_info_cache with network_info: [{"id": "07cc827a-876a-4432-a5fc-9ba8920dc5f9", "address": "fa:16:3e:71:e6:9f", "network": {"id": "06ab5813-9ad9-4021-9bdb-f2f02af8d73f", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1714653503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.175", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b2de35030a9484094e964ffc30a822d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap07cc827a-87", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1037.938031] env[63202]: DEBUG nova.compute.manager [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1037.940577] env[63202]: DEBUG nova.objects.base [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Object Instance<8437f856-a707-49c5-b8eb-5a22cdb990f8> lazy-loaded attributes: resources,numa_topology {{(pid=63202) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1037.971665] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "refresh_cache-3a65ddea-2d3d-45a1-a175-c8fad756793d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.971882] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquired lock "refresh_cache-3a65ddea-2d3d-45a1-a175-c8fad756793d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.971960] env[63202]: DEBUG nova.network.neutron [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1038.075950] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0390ccdc-d870-4fb0-b938-54799bd32d4a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.087472] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c47cee5a-b782-4e65-bc60-7e784f484ccf {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.121947] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecd27e22-bb97-4b21-b10d-e26d72ade153 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.130485] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb78a8c8-0a39-4e71-99f4-7caf742be8b4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.144960] env[63202]: DEBUG nova.compute.provider_tree [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1038.217266] env[63202]: DEBUG oslo_vmware.api [None req-7d0ec686-adfb-458e-8a0e-e648f13a89a8 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Task: {'id': task-1385807, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.164247} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.217530] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-7d0ec686-adfb-458e-8a0e-e648f13a89a8 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1038.217739] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-7d0ec686-adfb-458e-8a0e-e648f13a89a8 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Deleted contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1038.217933] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-7d0ec686-adfb-458e-8a0e-e648f13a89a8 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1038.218128] env[63202]: INFO nova.compute.manager [None req-7d0ec686-adfb-458e-8a0e-e648f13a89a8 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1038.218490] env[63202]: DEBUG oslo.service.loopingcall [None req-7d0ec686-adfb-458e-8a0e-e648f13a89a8 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1038.218793] env[63202]: DEBUG nova.compute.manager [-] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1038.218956] env[63202]: DEBUG nova.network.neutron [-] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1038.307842] env[63202]: DEBUG nova.network.neutron [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Successfully created port: 3cec3f36-0fb8-45c0-be6e-48cd6cf58aa3 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1038.401807] env[63202]: DEBUG oslo_concurrency.lockutils [req-393160ad-2ec6-486f-8bf7-771a816076ae req-39164185-9ec9-44fb-a722-e40bb9f9fa35 service nova] Releasing lock "refresh_cache-8437f856-a707-49c5-b8eb-5a22cdb990f8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1038.450319] env[63202]: DEBUG oslo_concurrency.lockutils [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquiring lock "8437f856-a707-49c5-b8eb-5a22cdb990f8" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.496380] env[63202]: DEBUG oslo_concurrency.lockutils [None req-43190153-e4ba-44d9-9717-98230b065546 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.496648] env[63202]: DEBUG oslo_concurrency.lockutils [None req-43190153-e4ba-44d9-9717-98230b065546 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.496853] env[63202]: DEBUG oslo_concurrency.lockutils [None req-43190153-e4ba-44d9-9717-98230b065546 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.497787] env[63202]: DEBUG oslo_concurrency.lockutils [None req-43190153-e4ba-44d9-9717-98230b065546 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.497994] env[63202]: DEBUG oslo_concurrency.lockutils [None req-43190153-e4ba-44d9-9717-98230b065546 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.500555] env[63202]: INFO nova.compute.manager [None req-43190153-e4ba-44d9-9717-98230b065546 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Terminating instance [ 1038.502429] env[63202]: DEBUG nova.compute.manager [None req-43190153-e4ba-44d9-9717-98230b065546 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1038.503036] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-43190153-e4ba-44d9-9717-98230b065546 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1038.503670] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f68a46c6-44f2-47fc-884b-c227ee2261fb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.512684] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-43190153-e4ba-44d9-9717-98230b065546 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1038.513202] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b929a9e0-edcd-4ee7-8476-20f7a14686db {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.529708] env[63202]: DEBUG nova.network.neutron [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1038.648353] env[63202]: DEBUG nova.scheduler.client.report [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1038.699776] env[63202]: DEBUG nova.network.neutron [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Updating instance_info_cache with network_info: [{"id": "3bfc4703-af3e-41fe-a35a-a5a8fadf1aef", "address": "fa:16:3e:03:78:69", "network": {"id": "b1ea2320-3801-43fe-b566-4f69db81329e", "bridge": "br-int", "label": "tempest-ServersTestJSON-618582355-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e972faeaedc6468aab7e7cfee88a477b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4223acd2-30f7-440e-b975-60b30d931694", "external-id": "nsx-vlan-transportzone-647", "segmentation_id": 647, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3bfc4703-af", "ovs_interfaceid": "3bfc4703-af3e-41fe-a35a-a5a8fadf1aef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1038.707627] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-43190153-e4ba-44d9-9717-98230b065546 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1038.707914] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-43190153-e4ba-44d9-9717-98230b065546 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Deleting contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1038.708138] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-43190153-e4ba-44d9-9717-98230b065546 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Deleting the datastore file [datastore1] 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1038.708424] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-85d69ced-95e4-4acf-99af-4f9442ff0fa8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.716174] env[63202]: DEBUG oslo_vmware.api [None req-43190153-e4ba-44d9-9717-98230b065546 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 1038.716174] env[63202]: value = "task-1385809" [ 1038.716174] env[63202]: _type = "Task" [ 1038.716174] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.725574] env[63202]: DEBUG oslo_vmware.api [None req-43190153-e4ba-44d9-9717-98230b065546 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385809, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.952382] env[63202]: DEBUG nova.compute.manager [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1038.989940] env[63202]: DEBUG nova.virt.hardware [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1038.990221] env[63202]: DEBUG nova.virt.hardware [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1038.990385] env[63202]: DEBUG nova.virt.hardware [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1038.990569] env[63202]: DEBUG nova.virt.hardware [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1038.990721] env[63202]: DEBUG nova.virt.hardware [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1038.990874] env[63202]: DEBUG nova.virt.hardware [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1038.991153] env[63202]: DEBUG nova.virt.hardware [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1038.991341] env[63202]: DEBUG nova.virt.hardware [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1038.991516] env[63202]: DEBUG nova.virt.hardware [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1038.991682] env[63202]: DEBUG nova.virt.hardware [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1038.991856] env[63202]: DEBUG nova.virt.hardware [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1038.992750] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02b5d0f8-8c70-4857-8b75-223912ec44b6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.000716] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cce4d040-cde9-4d55-a757-c0220a8aa73c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.035153] env[63202]: DEBUG oslo_concurrency.lockutils [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Acquiring lock "1123a12e-5218-415d-b286-2f005fe57b29" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.035392] env[63202]: DEBUG oslo_concurrency.lockutils [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Lock "1123a12e-5218-415d-b286-2f005fe57b29" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.071123] env[63202]: DEBUG nova.network.neutron [-] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1039.086177] env[63202]: DEBUG nova.compute.manager [req-e7768968-5d2f-4319-8d93-c74aaf66394f req-010b8a97-cb30-4d27-8c64-5800d76a9871 service nova] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Received event network-vif-plugged-3bfc4703-af3e-41fe-a35a-a5a8fadf1aef {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1039.086177] env[63202]: DEBUG oslo_concurrency.lockutils [req-e7768968-5d2f-4319-8d93-c74aaf66394f req-010b8a97-cb30-4d27-8c64-5800d76a9871 service nova] Acquiring lock "3a65ddea-2d3d-45a1-a175-c8fad756793d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.086177] env[63202]: DEBUG oslo_concurrency.lockutils [req-e7768968-5d2f-4319-8d93-c74aaf66394f req-010b8a97-cb30-4d27-8c64-5800d76a9871 service nova] Lock "3a65ddea-2d3d-45a1-a175-c8fad756793d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.086177] env[63202]: DEBUG oslo_concurrency.lockutils [req-e7768968-5d2f-4319-8d93-c74aaf66394f req-010b8a97-cb30-4d27-8c64-5800d76a9871 service nova] Lock "3a65ddea-2d3d-45a1-a175-c8fad756793d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.086177] env[63202]: DEBUG nova.compute.manager [req-e7768968-5d2f-4319-8d93-c74aaf66394f req-010b8a97-cb30-4d27-8c64-5800d76a9871 service nova] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] No waiting events found dispatching network-vif-plugged-3bfc4703-af3e-41fe-a35a-a5a8fadf1aef {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1039.086833] env[63202]: WARNING nova.compute.manager [req-e7768968-5d2f-4319-8d93-c74aaf66394f req-010b8a97-cb30-4d27-8c64-5800d76a9871 service nova] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Received unexpected event network-vif-plugged-3bfc4703-af3e-41fe-a35a-a5a8fadf1aef for instance with vm_state building and task_state spawning. [ 1039.086833] env[63202]: DEBUG nova.compute.manager [req-e7768968-5d2f-4319-8d93-c74aaf66394f req-010b8a97-cb30-4d27-8c64-5800d76a9871 service nova] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Received event network-changed-3bfc4703-af3e-41fe-a35a-a5a8fadf1aef {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1039.086833] env[63202]: DEBUG nova.compute.manager [req-e7768968-5d2f-4319-8d93-c74aaf66394f req-010b8a97-cb30-4d27-8c64-5800d76a9871 service nova] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Refreshing instance network info cache due to event network-changed-3bfc4703-af3e-41fe-a35a-a5a8fadf1aef. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1039.086936] env[63202]: DEBUG oslo_concurrency.lockutils [req-e7768968-5d2f-4319-8d93-c74aaf66394f req-010b8a97-cb30-4d27-8c64-5800d76a9871 service nova] Acquiring lock "refresh_cache-3a65ddea-2d3d-45a1-a175-c8fad756793d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1039.140633] env[63202]: INFO nova.compute.manager [None req-785be6b2-df40-4037-b7c3-91924a04c311 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Rebuilding instance [ 1039.154030] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.226s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.184126] env[63202]: DEBUG nova.compute.manager [None req-785be6b2-df40-4037-b7c3-91924a04c311 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1039.185034] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54f4f220-ba55-48e0-a4a5-30629c2dc694 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.202691] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Releasing lock "refresh_cache-3a65ddea-2d3d-45a1-a175-c8fad756793d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1039.203665] env[63202]: DEBUG nova.compute.manager [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Instance network_info: |[{"id": "3bfc4703-af3e-41fe-a35a-a5a8fadf1aef", "address": "fa:16:3e:03:78:69", "network": {"id": "b1ea2320-3801-43fe-b566-4f69db81329e", "bridge": "br-int", "label": "tempest-ServersTestJSON-618582355-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e972faeaedc6468aab7e7cfee88a477b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4223acd2-30f7-440e-b975-60b30d931694", "external-id": "nsx-vlan-transportzone-647", "segmentation_id": 647, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3bfc4703-af", "ovs_interfaceid": "3bfc4703-af3e-41fe-a35a-a5a8fadf1aef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1039.203665] env[63202]: DEBUG oslo_concurrency.lockutils [req-e7768968-5d2f-4319-8d93-c74aaf66394f req-010b8a97-cb30-4d27-8c64-5800d76a9871 service nova] Acquired lock "refresh_cache-3a65ddea-2d3d-45a1-a175-c8fad756793d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.203871] env[63202]: DEBUG nova.network.neutron [req-e7768968-5d2f-4319-8d93-c74aaf66394f req-010b8a97-cb30-4d27-8c64-5800d76a9871 service nova] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Refreshing network info cache for port 3bfc4703-af3e-41fe-a35a-a5a8fadf1aef {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1039.204858] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:03:78:69', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4223acd2-30f7-440e-b975-60b30d931694', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3bfc4703-af3e-41fe-a35a-a5a8fadf1aef', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1039.212535] env[63202]: DEBUG oslo.service.loopingcall [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1039.212971] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1039.213392] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-11d7d639-43dd-4ff4-aea5-2fd510221e16 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.241343] env[63202]: DEBUG oslo_vmware.api [None req-43190153-e4ba-44d9-9717-98230b065546 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385809, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.251361} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.242603] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-43190153-e4ba-44d9-9717-98230b065546 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1039.242800] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-43190153-e4ba-44d9-9717-98230b065546 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Deleted contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1039.242974] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-43190153-e4ba-44d9-9717-98230b065546 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1039.243164] env[63202]: INFO nova.compute.manager [None req-43190153-e4ba-44d9-9717-98230b065546 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Took 0.74 seconds to destroy the instance on the hypervisor. [ 1039.243407] env[63202]: DEBUG oslo.service.loopingcall [None req-43190153-e4ba-44d9-9717-98230b065546 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1039.243584] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1039.243584] env[63202]: value = "task-1385810" [ 1039.243584] env[63202]: _type = "Task" [ 1039.243584] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.243766] env[63202]: DEBUG nova.compute.manager [-] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1039.243908] env[63202]: DEBUG nova.network.neutron [-] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1039.257802] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385810, 'name': CreateVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.537842] env[63202]: DEBUG nova.compute.manager [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1039.573914] env[63202]: INFO nova.compute.manager [-] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Took 1.35 seconds to deallocate network for instance. [ 1039.663807] env[63202]: DEBUG oslo_concurrency.lockutils [None req-4d5d1ebd-110d-4dc9-9f25-39163bab49c7 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lock "8437f856-a707-49c5-b8eb-5a22cdb990f8" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 20.870s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.664499] env[63202]: DEBUG oslo_concurrency.lockutils [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lock "8437f856-a707-49c5-b8eb-5a22cdb990f8" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.215s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.664829] env[63202]: INFO nova.compute.manager [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Unshelving [ 1039.696770] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-785be6b2-df40-4037-b7c3-91924a04c311 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1039.697135] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8cf02ecf-62ef-40ef-94c0-a8ec15a276d1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.705251] env[63202]: DEBUG oslo_vmware.api [None req-785be6b2-df40-4037-b7c3-91924a04c311 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Waiting for the task: (returnval){ [ 1039.705251] env[63202]: value = "task-1385811" [ 1039.705251] env[63202]: _type = "Task" [ 1039.705251] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.716158] env[63202]: DEBUG oslo_vmware.api [None req-785be6b2-df40-4037-b7c3-91924a04c311 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Task: {'id': task-1385811, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.757848] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385810, 'name': CreateVM_Task, 'duration_secs': 0.346829} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.758036] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1039.758723] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1039.758905] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.759250] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1039.759531] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a384d73e-683b-4568-984f-6821fb1838f9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.766691] env[63202]: DEBUG oslo_vmware.api [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 1039.766691] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52f42709-6c57-e2bd-aa18-066365a7df02" [ 1039.766691] env[63202]: _type = "Task" [ 1039.766691] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.778725] env[63202]: DEBUG oslo_vmware.api [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52f42709-6c57-e2bd-aa18-066365a7df02, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.917936] env[63202]: DEBUG nova.compute.manager [req-56f81b98-9c87-4fea-a3da-2f242e65e7e3 req-bda4bd60-0b71-4ca3-830d-9f720d7d3b7e service nova] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Received event network-vif-plugged-3cec3f36-0fb8-45c0-be6e-48cd6cf58aa3 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1039.918186] env[63202]: DEBUG oslo_concurrency.lockutils [req-56f81b98-9c87-4fea-a3da-2f242e65e7e3 req-bda4bd60-0b71-4ca3-830d-9f720d7d3b7e service nova] Acquiring lock "e1e2e8a4-80aa-45eb-a90e-d20cba2943ee-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.918393] env[63202]: DEBUG oslo_concurrency.lockutils [req-56f81b98-9c87-4fea-a3da-2f242e65e7e3 req-bda4bd60-0b71-4ca3-830d-9f720d7d3b7e service nova] Lock "e1e2e8a4-80aa-45eb-a90e-d20cba2943ee-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.918558] env[63202]: DEBUG oslo_concurrency.lockutils [req-56f81b98-9c87-4fea-a3da-2f242e65e7e3 req-bda4bd60-0b71-4ca3-830d-9f720d7d3b7e service nova] Lock "e1e2e8a4-80aa-45eb-a90e-d20cba2943ee-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.918731] env[63202]: DEBUG nova.compute.manager [req-56f81b98-9c87-4fea-a3da-2f242e65e7e3 req-bda4bd60-0b71-4ca3-830d-9f720d7d3b7e service nova] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] No waiting events found dispatching network-vif-plugged-3cec3f36-0fb8-45c0-be6e-48cd6cf58aa3 {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1039.918900] env[63202]: WARNING nova.compute.manager [req-56f81b98-9c87-4fea-a3da-2f242e65e7e3 req-bda4bd60-0b71-4ca3-830d-9f720d7d3b7e service nova] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Received unexpected event network-vif-plugged-3cec3f36-0fb8-45c0-be6e-48cd6cf58aa3 for instance with vm_state building and task_state spawning. [ 1039.934363] env[63202]: DEBUG nova.network.neutron [req-e7768968-5d2f-4319-8d93-c74aaf66394f req-010b8a97-cb30-4d27-8c64-5800d76a9871 service nova] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Updated VIF entry in instance network info cache for port 3bfc4703-af3e-41fe-a35a-a5a8fadf1aef. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1039.934748] env[63202]: DEBUG nova.network.neutron [req-e7768968-5d2f-4319-8d93-c74aaf66394f req-010b8a97-cb30-4d27-8c64-5800d76a9871 service nova] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Updating instance_info_cache with network_info: [{"id": "3bfc4703-af3e-41fe-a35a-a5a8fadf1aef", "address": "fa:16:3e:03:78:69", "network": {"id": "b1ea2320-3801-43fe-b566-4f69db81329e", "bridge": "br-int", "label": "tempest-ServersTestJSON-618582355-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e972faeaedc6468aab7e7cfee88a477b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4223acd2-30f7-440e-b975-60b30d931694", "external-id": "nsx-vlan-transportzone-647", "segmentation_id": 647, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3bfc4703-af", "ovs_interfaceid": "3bfc4703-af3e-41fe-a35a-a5a8fadf1aef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.018463] env[63202]: DEBUG nova.network.neutron [-] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.061757] env[63202]: DEBUG oslo_concurrency.lockutils [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.062043] env[63202]: DEBUG oslo_concurrency.lockutils [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.063522] env[63202]: INFO nova.compute.claims [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1040.084560] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7d0ec686-adfb-458e-8a0e-e648f13a89a8 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.144198] env[63202]: DEBUG nova.network.neutron [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Successfully updated port: 3cec3f36-0fb8-45c0-be6e-48cd6cf58aa3 {{(pid=63202) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1040.215457] env[63202]: DEBUG oslo_vmware.api [None req-785be6b2-df40-4037-b7c3-91924a04c311 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Task: {'id': task-1385811, 'name': PowerOffVM_Task, 'duration_secs': 0.187637} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.215723] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-785be6b2-df40-4037-b7c3-91924a04c311 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1040.216528] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-785be6b2-df40-4037-b7c3-91924a04c311 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1040.216795] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0703095b-020b-4c6a-99e0-cac0dde30a62 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.224857] env[63202]: DEBUG oslo_vmware.api [None req-785be6b2-df40-4037-b7c3-91924a04c311 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Waiting for the task: (returnval){ [ 1040.224857] env[63202]: value = "task-1385812" [ 1040.224857] env[63202]: _type = "Task" [ 1040.224857] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.232966] env[63202]: DEBUG oslo_vmware.api [None req-785be6b2-df40-4037-b7c3-91924a04c311 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Task: {'id': task-1385812, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.278013] env[63202]: DEBUG oslo_vmware.api [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52f42709-6c57-e2bd-aa18-066365a7df02, 'name': SearchDatastore_Task, 'duration_secs': 0.010804} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.278331] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1040.278633] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1040.278958] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1040.279184] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.279465] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1040.279807] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-01c9e156-9a14-4c4c-8b26-a1bbc7993a56 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.288625] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1040.288800] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1040.289544] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-61e5788a-f717-4df5-a09e-c149a17c8f93 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.294511] env[63202]: DEBUG oslo_vmware.api [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 1040.294511] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52569a52-8035-5421-5a1d-9a10cd664cae" [ 1040.294511] env[63202]: _type = "Task" [ 1040.294511] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.301789] env[63202]: DEBUG oslo_vmware.api [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52569a52-8035-5421-5a1d-9a10cd664cae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.437113] env[63202]: DEBUG oslo_concurrency.lockutils [req-e7768968-5d2f-4319-8d93-c74aaf66394f req-010b8a97-cb30-4d27-8c64-5800d76a9871 service nova] Releasing lock "refresh_cache-3a65ddea-2d3d-45a1-a175-c8fad756793d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1040.437382] env[63202]: DEBUG nova.compute.manager [req-e7768968-5d2f-4319-8d93-c74aaf66394f req-010b8a97-cb30-4d27-8c64-5800d76a9871 service nova] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Received event network-vif-deleted-ea5e41d2-19a6-4899-9bfa-c9b18de354e6 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1040.437565] env[63202]: INFO nova.compute.manager [req-e7768968-5d2f-4319-8d93-c74aaf66394f req-010b8a97-cb30-4d27-8c64-5800d76a9871 service nova] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Neutron deleted interface ea5e41d2-19a6-4899-9bfa-c9b18de354e6; detaching it from the instance and deleting it from the info cache [ 1040.437735] env[63202]: DEBUG nova.network.neutron [req-e7768968-5d2f-4319-8d93-c74aaf66394f req-010b8a97-cb30-4d27-8c64-5800d76a9871 service nova] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.521501] env[63202]: INFO nova.compute.manager [-] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Took 1.28 seconds to deallocate network for instance. [ 1040.645387] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Acquiring lock "refresh_cache-e1e2e8a4-80aa-45eb-a90e-d20cba2943ee" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1040.645698] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Acquired lock "refresh_cache-e1e2e8a4-80aa-45eb-a90e-d20cba2943ee" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.645743] env[63202]: DEBUG nova.network.neutron [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1040.674670] env[63202]: DEBUG nova.compute.utils [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1040.735440] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-785be6b2-df40-4037-b7c3-91924a04c311 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] VM already powered off {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1040.735605] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-785be6b2-df40-4037-b7c3-91924a04c311 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Volume detach. Driver type: vmdk {{(pid=63202) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1040.735798] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-785be6b2-df40-4037-b7c3-91924a04c311 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294182', 'volume_id': '6518fe7c-9ed9-49fb-b419-6c72d538c63d', 'name': 'volume-6518fe7c-9ed9-49fb-b419-6c72d538c63d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0d7fd297-37db-4876-bb91-5bb0b7dc335e', 'attached_at': '', 'detached_at': '', 'volume_id': '6518fe7c-9ed9-49fb-b419-6c72d538c63d', 'serial': '6518fe7c-9ed9-49fb-b419-6c72d538c63d'} {{(pid=63202) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1040.736592] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e9e7680-72ea-4dd2-9517-2a21461590fd {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.756621] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb6142c6-bda5-4d55-b326-644d9e102cad {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.471875] env[63202]: DEBUG oslo_concurrency.lockutils [None req-43190153-e4ba-44d9-9717-98230b065546 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.476369] env[63202]: INFO nova.virt.block_device [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Booting with volume 81b406ab-e721-4271-b298-75cded690f20 at /dev/sdb [ 1041.478564] env[63202]: DEBUG nova.compute.manager [req-f1325429-ef33-46ea-904f-f5251b3016a8 req-1a5b4fb2-38c4-4da2-a260-c7442c3c7506 service nova] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Received event network-vif-deleted-5f6faad3-ee63-4318-abef-e6442d090681 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1041.481943] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b30eb0d0-a9d8-4727-8820-f7a0b5720a09 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.493622] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-814a1965-e42c-4708-aafc-011a1d2cfa3e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.497154] env[63202]: DEBUG oslo_vmware.api [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52569a52-8035-5421-5a1d-9a10cd664cae, 'name': SearchDatastore_Task, 'duration_secs': 0.007427} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.501797] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-681b8126-935c-46a5-a2d5-766a966b5a0c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.506715] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bd9a6c7-8351-4048-a439-74bbcc9b1298 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.530707] env[63202]: DEBUG nova.network.neutron [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1041.534618] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1b7b381-6093-405c-a257-f6e593344877 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.537251] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d95836d2-ac2a-4558-8639-7148314c7aba {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.540069] env[63202]: DEBUG oslo_vmware.api [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 1041.540069] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]5276ff8a-7898-be56-c538-097d8123e10e" [ 1041.540069] env[63202]: _type = "Task" [ 1041.540069] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.554221] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-785be6b2-df40-4037-b7c3-91924a04c311 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] The volume has not been displaced from its original location: [datastore2] volume-6518fe7c-9ed9-49fb-b419-6c72d538c63d/volume-6518fe7c-9ed9-49fb-b419-6c72d538c63d.vmdk. No consolidation needed. {{(pid=63202) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1041.559555] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-785be6b2-df40-4037-b7c3-91924a04c311 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Reconfiguring VM instance instance-0000005b to detach disk 2000 {{(pid=63202) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1041.570699] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7fc728be-b7f9-46f6-8d4d-550669ff4912 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.583519] env[63202]: DEBUG nova.compute.manager [req-e7768968-5d2f-4319-8d93-c74aaf66394f req-010b8a97-cb30-4d27-8c64-5800d76a9871 service nova] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Detach interface failed, port_id=ea5e41d2-19a6-4899-9bfa-c9b18de354e6, reason: Instance c2b6f3a4-7414-4fc4-893b-9b613fc5e381 could not be found. {{(pid=63202) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1041.590101] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-942a6b17-f2dc-4f85-ba7b-a0108c4e4ae4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.600929] env[63202]: DEBUG oslo_vmware.api [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]5276ff8a-7898-be56-c538-097d8123e10e, 'name': SearchDatastore_Task, 'duration_secs': 0.00964} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.603996] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1041.604282] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] 3a65ddea-2d3d-45a1-a175-c8fad756793d/3a65ddea-2d3d-45a1-a175-c8fad756793d.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1041.604992] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d3ff9308-6c5f-47b7-aabb-cf86c9749139 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.609600] env[63202]: DEBUG oslo_vmware.api [None req-785be6b2-df40-4037-b7c3-91924a04c311 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Waiting for the task: (returnval){ [ 1041.609600] env[63202]: value = "task-1385813" [ 1041.609600] env[63202]: _type = "Task" [ 1041.609600] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.619536] env[63202]: DEBUG oslo_vmware.api [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 1041.619536] env[63202]: value = "task-1385814" [ 1041.619536] env[63202]: _type = "Task" [ 1041.619536] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.619754] env[63202]: DEBUG oslo_vmware.api [None req-785be6b2-df40-4037-b7c3-91924a04c311 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Task: {'id': task-1385813, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.635229] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d63138aa-7f21-4d07-a987-57e04d6fba29 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.638581] env[63202]: DEBUG oslo_vmware.api [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385814, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.644195] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2af30ca-5165-4599-ae07-67da8ff9b3f9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.679792] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ece2c8b-a9ae-43f3-abbd-9332c3b45c6a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.682818] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e1bbbd3-08cb-472d-9e29-f6a74d766353 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.692248] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fe4d622-8a34-44c9-92de-b1cb8d19ef7a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.694866] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc72cf67-c280-434b-b80f-68946493b338 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.731018] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00cdb425-d160-4565-afd5-3388bf03ca49 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.733611] env[63202]: DEBUG nova.virt.block_device [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Updating existing volume attachment record: fdfb419e-c6e2-4a9a-92fd-a47ad5e49c30 {{(pid=63202) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1041.743093] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee209abb-ca3a-48ba-acb3-c429ce99a2c8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.757461] env[63202]: DEBUG nova.compute.provider_tree [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1041.803283] env[63202]: DEBUG nova.network.neutron [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Updating instance_info_cache with network_info: [{"id": "3cec3f36-0fb8-45c0-be6e-48cd6cf58aa3", "address": "fa:16:3e:ac:21:28", "network": {"id": "345538f4-e791-4fc6-9719-237c11f1382a", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-449651008-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a9e939cdf4a94a8aa0107a5761771c2c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa09e855-8af1-419b-b78d-8ffcc94b1bfb", "external-id": "nsx-vlan-transportzone-901", "segmentation_id": 901, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cec3f36-0f", "ovs_interfaceid": "3cec3f36-0fb8-45c0-be6e-48cd6cf58aa3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1041.951054] env[63202]: DEBUG nova.compute.manager [req-0bdb0632-767c-4159-bb5a-128f4234a477 req-87be931e-cc37-4241-bbf6-667c089f63dd service nova] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Received event network-changed-3cec3f36-0fb8-45c0-be6e-48cd6cf58aa3 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1041.951431] env[63202]: DEBUG nova.compute.manager [req-0bdb0632-767c-4159-bb5a-128f4234a477 req-87be931e-cc37-4241-bbf6-667c089f63dd service nova] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Refreshing instance network info cache due to event network-changed-3cec3f36-0fb8-45c0-be6e-48cd6cf58aa3. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1041.951431] env[63202]: DEBUG oslo_concurrency.lockutils [req-0bdb0632-767c-4159-bb5a-128f4234a477 req-87be931e-cc37-4241-bbf6-667c089f63dd service nova] Acquiring lock "refresh_cache-e1e2e8a4-80aa-45eb-a90e-d20cba2943ee" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.119938] env[63202]: DEBUG oslo_vmware.api [None req-785be6b2-df40-4037-b7c3-91924a04c311 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Task: {'id': task-1385813, 'name': ReconfigVM_Task, 'duration_secs': 0.235254} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.120353] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-785be6b2-df40-4037-b7c3-91924a04c311 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Reconfigured VM instance instance-0000005b to detach disk 2000 {{(pid=63202) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1042.127877] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-35c6517a-4860-4a33-9611-3157f6c741c4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.143353] env[63202]: DEBUG oslo_vmware.api [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385814, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.458267} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.144560] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] 3a65ddea-2d3d-45a1-a175-c8fad756793d/3a65ddea-2d3d-45a1-a175-c8fad756793d.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1042.144783] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1042.145096] env[63202]: DEBUG oslo_vmware.api [None req-785be6b2-df40-4037-b7c3-91924a04c311 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Waiting for the task: (returnval){ [ 1042.145096] env[63202]: value = "task-1385818" [ 1042.145096] env[63202]: _type = "Task" [ 1042.145096] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.145293] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0eef3e27-262f-40d3-9fce-4bf89361290c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.156008] env[63202]: DEBUG oslo_vmware.api [None req-785be6b2-df40-4037-b7c3-91924a04c311 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Task: {'id': task-1385818, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.157765] env[63202]: DEBUG oslo_vmware.api [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 1042.157765] env[63202]: value = "task-1385819" [ 1042.157765] env[63202]: _type = "Task" [ 1042.157765] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.168889] env[63202]: DEBUG oslo_vmware.api [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385819, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.260875] env[63202]: DEBUG nova.scheduler.client.report [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1042.305858] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Releasing lock "refresh_cache-e1e2e8a4-80aa-45eb-a90e-d20cba2943ee" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1042.306204] env[63202]: DEBUG nova.compute.manager [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Instance network_info: |[{"id": "3cec3f36-0fb8-45c0-be6e-48cd6cf58aa3", "address": "fa:16:3e:ac:21:28", "network": {"id": "345538f4-e791-4fc6-9719-237c11f1382a", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-449651008-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a9e939cdf4a94a8aa0107a5761771c2c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa09e855-8af1-419b-b78d-8ffcc94b1bfb", "external-id": "nsx-vlan-transportzone-901", "segmentation_id": 901, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cec3f36-0f", "ovs_interfaceid": "3cec3f36-0fb8-45c0-be6e-48cd6cf58aa3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1042.306532] env[63202]: DEBUG oslo_concurrency.lockutils [req-0bdb0632-767c-4159-bb5a-128f4234a477 req-87be931e-cc37-4241-bbf6-667c089f63dd service nova] Acquired lock "refresh_cache-e1e2e8a4-80aa-45eb-a90e-d20cba2943ee" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.306737] env[63202]: DEBUG nova.network.neutron [req-0bdb0632-767c-4159-bb5a-128f4234a477 req-87be931e-cc37-4241-bbf6-667c089f63dd service nova] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Refreshing network info cache for port 3cec3f36-0fb8-45c0-be6e-48cd6cf58aa3 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1042.307922] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ac:21:28', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'aa09e855-8af1-419b-b78d-8ffcc94b1bfb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3cec3f36-0fb8-45c0-be6e-48cd6cf58aa3', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1042.316560] env[63202]: DEBUG oslo.service.loopingcall [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1042.319527] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1042.320053] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4802f87a-a7da-4803-a9c5-88504a71624f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.340586] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1042.340586] env[63202]: value = "task-1385820" [ 1042.340586] env[63202]: _type = "Task" [ 1042.340586] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.349149] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385820, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.546777] env[63202]: DEBUG nova.network.neutron [req-0bdb0632-767c-4159-bb5a-128f4234a477 req-87be931e-cc37-4241-bbf6-667c089f63dd service nova] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Updated VIF entry in instance network info cache for port 3cec3f36-0fb8-45c0-be6e-48cd6cf58aa3. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1042.547287] env[63202]: DEBUG nova.network.neutron [req-0bdb0632-767c-4159-bb5a-128f4234a477 req-87be931e-cc37-4241-bbf6-667c089f63dd service nova] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Updating instance_info_cache with network_info: [{"id": "3cec3f36-0fb8-45c0-be6e-48cd6cf58aa3", "address": "fa:16:3e:ac:21:28", "network": {"id": "345538f4-e791-4fc6-9719-237c11f1382a", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-449651008-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a9e939cdf4a94a8aa0107a5761771c2c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa09e855-8af1-419b-b78d-8ffcc94b1bfb", "external-id": "nsx-vlan-transportzone-901", "segmentation_id": 901, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cec3f36-0f", "ovs_interfaceid": "3cec3f36-0fb8-45c0-be6e-48cd6cf58aa3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.659802] env[63202]: DEBUG oslo_vmware.api [None req-785be6b2-df40-4037-b7c3-91924a04c311 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Task: {'id': task-1385818, 'name': ReconfigVM_Task, 'duration_secs': 0.124852} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.662927] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-785be6b2-df40-4037-b7c3-91924a04c311 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294182', 'volume_id': '6518fe7c-9ed9-49fb-b419-6c72d538c63d', 'name': 'volume-6518fe7c-9ed9-49fb-b419-6c72d538c63d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0d7fd297-37db-4876-bb91-5bb0b7dc335e', 'attached_at': '', 'detached_at': '', 'volume_id': '6518fe7c-9ed9-49fb-b419-6c72d538c63d', 'serial': '6518fe7c-9ed9-49fb-b419-6c72d538c63d'} {{(pid=63202) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1042.663256] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-785be6b2-df40-4037-b7c3-91924a04c311 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1042.664017] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e92ac55c-88bb-49b8-a66b-2077610c3a60 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.671758] env[63202]: DEBUG oslo_vmware.api [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385819, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065418} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.673711] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1042.674018] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-785be6b2-df40-4037-b7c3-91924a04c311 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1042.674724] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48898543-6bc4-44d9-8da3-49193562b359 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.677294] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-afa18efd-cff8-40e6-9e96-a908fecdf85f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.701495] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Reconfiguring VM instance instance-00000061 to attach disk [datastore1] 3a65ddea-2d3d-45a1-a175-c8fad756793d/3a65ddea-2d3d-45a1-a175-c8fad756793d.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1042.701843] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7e94f436-cc5a-4b48-a44b-500333fcf50a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.726406] env[63202]: DEBUG oslo_vmware.api [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 1042.726406] env[63202]: value = "task-1385822" [ 1042.726406] env[63202]: _type = "Task" [ 1042.726406] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.736600] env[63202]: DEBUG oslo_vmware.api [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385822, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.754047] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-785be6b2-df40-4037-b7c3-91924a04c311 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1042.754293] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-785be6b2-df40-4037-b7c3-91924a04c311 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Deleting contents of the VM from datastore datastore2 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1042.754482] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-785be6b2-df40-4037-b7c3-91924a04c311 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Deleting the datastore file [datastore2] 0d7fd297-37db-4876-bb91-5bb0b7dc335e {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1042.754763] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1dcd026d-fb97-4f22-8053-8e51f465275d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.763775] env[63202]: DEBUG oslo_vmware.api [None req-785be6b2-df40-4037-b7c3-91924a04c311 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Waiting for the task: (returnval){ [ 1042.763775] env[63202]: value = "task-1385823" [ 1042.763775] env[63202]: _type = "Task" [ 1042.763775] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.767466] env[63202]: DEBUG oslo_concurrency.lockutils [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.705s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.767965] env[63202]: DEBUG nova.compute.manager [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1042.771494] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7d0ec686-adfb-458e-8a0e-e648f13a89a8 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.687s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.771746] env[63202]: DEBUG nova.objects.instance [None req-7d0ec686-adfb-458e-8a0e-e648f13a89a8 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lazy-loading 'resources' on Instance uuid c2b6f3a4-7414-4fc4-893b-9b613fc5e381 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1042.779482] env[63202]: DEBUG oslo_vmware.api [None req-785be6b2-df40-4037-b7c3-91924a04c311 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Task: {'id': task-1385823, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.852846] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385820, 'name': CreateVM_Task, 'duration_secs': 0.476704} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.853051] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1042.853803] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.853979] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.854336] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1042.854607] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-08020f9d-9059-419e-98e8-cdf10ea6141f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.860218] env[63202]: DEBUG oslo_vmware.api [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Waiting for the task: (returnval){ [ 1042.860218] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52d3ee65-998b-04d6-7198-73b6d3097e39" [ 1042.860218] env[63202]: _type = "Task" [ 1042.860218] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.868191] env[63202]: DEBUG oslo_vmware.api [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52d3ee65-998b-04d6-7198-73b6d3097e39, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.050902] env[63202]: DEBUG oslo_concurrency.lockutils [req-0bdb0632-767c-4159-bb5a-128f4234a477 req-87be931e-cc37-4241-bbf6-667c089f63dd service nova] Releasing lock "refresh_cache-e1e2e8a4-80aa-45eb-a90e-d20cba2943ee" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.236604] env[63202]: DEBUG oslo_vmware.api [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385822, 'name': ReconfigVM_Task, 'duration_secs': 0.30732} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.236887] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Reconfigured VM instance instance-00000061 to attach disk [datastore1] 3a65ddea-2d3d-45a1-a175-c8fad756793d/3a65ddea-2d3d-45a1-a175-c8fad756793d.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1043.237512] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d8ad9828-27cd-406e-992d-d8182781bb34 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.244453] env[63202]: DEBUG oslo_vmware.api [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 1043.244453] env[63202]: value = "task-1385824" [ 1043.244453] env[63202]: _type = "Task" [ 1043.244453] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.251616] env[63202]: DEBUG oslo_vmware.api [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385824, 'name': Rename_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.275249] env[63202]: DEBUG nova.compute.utils [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1043.279048] env[63202]: DEBUG oslo_vmware.api [None req-785be6b2-df40-4037-b7c3-91924a04c311 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Task: {'id': task-1385823, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.093229} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.280394] env[63202]: DEBUG nova.compute.manager [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1043.280394] env[63202]: DEBUG nova.network.neutron [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1043.282428] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-785be6b2-df40-4037-b7c3-91924a04c311 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1043.283221] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-785be6b2-df40-4037-b7c3-91924a04c311 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Deleted contents of the VM from datastore datastore2 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1043.283221] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-785be6b2-df40-4037-b7c3-91924a04c311 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1043.334052] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-785be6b2-df40-4037-b7c3-91924a04c311 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Volume detach. Driver type: vmdk {{(pid=63202) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1043.334431] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-01c62db8-eba6-4053-9237-9ebeafa30f9b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.345887] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c15b206-25ec-4de2-8777-e3d69693cc36 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.367318] env[63202]: DEBUG nova.policy [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '662e072d7aa44efb994b35b2a6656dec', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9c1a31c777da417c96b2552ca50dc3a7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 1043.380571] env[63202]: ERROR nova.compute.manager [None req-785be6b2-df40-4037-b7c3-91924a04c311 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Failed to detach volume 6518fe7c-9ed9-49fb-b419-6c72d538c63d from /dev/sda: nova.exception.InstanceNotFound: Instance 0d7fd297-37db-4876-bb91-5bb0b7dc335e could not be found. [ 1043.380571] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Traceback (most recent call last): [ 1043.380571] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 1043.380571] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] self.driver.rebuild(**kwargs) [ 1043.380571] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] File "/opt/stack/nova/nova/virt/driver.py", line 493, in rebuild [ 1043.380571] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] raise NotImplementedError() [ 1043.380571] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] NotImplementedError [ 1043.380571] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] [ 1043.380571] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] During handling of the above exception, another exception occurred: [ 1043.380571] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] [ 1043.380571] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Traceback (most recent call last): [ 1043.380571] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 1043.380571] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] self.driver.detach_volume(context, old_connection_info, [ 1043.381305] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 559, in detach_volume [ 1043.381305] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] return self._volumeops.detach_volume(connection_info, instance) [ 1043.381305] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 1043.381305] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] self._detach_volume_vmdk(connection_info, instance) [ 1043.381305] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 1043.381305] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 1043.381305] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 1043.381305] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] stable_ref.fetch_moref(session) [ 1043.381305] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 1043.381305] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] raise exception.InstanceNotFound(instance_id=self._uuid) [ 1043.381305] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] nova.exception.InstanceNotFound: Instance 0d7fd297-37db-4876-bb91-5bb0b7dc335e could not be found. [ 1043.381305] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] [ 1043.388212] env[63202]: DEBUG oslo_vmware.api [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52d3ee65-998b-04d6-7198-73b6d3097e39, 'name': SearchDatastore_Task, 'duration_secs': 0.009168} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.388673] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.388902] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1043.389166] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.389292] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.389502] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1043.389720] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7091dd52-c85b-4422-8934-261c7ed84aa4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.398909] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1043.399864] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1043.402142] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5fddf29c-aeb1-47ad-87c8-227df6bb4331 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.408538] env[63202]: DEBUG oslo_vmware.api [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Waiting for the task: (returnval){ [ 1043.408538] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]524fab16-02de-f3ad-5334-147bf0261254" [ 1043.408538] env[63202]: _type = "Task" [ 1043.408538] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.415162] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3213e696-de28-41cb-a439-bd1d8364b7af {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.420503] env[63202]: DEBUG oslo_vmware.api [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]524fab16-02de-f3ad-5334-147bf0261254, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.427700] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90d882d6-6372-41d4-8396-682664b7f12d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.457451] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1090c0d6-5edc-4f59-b86c-52fb7266b147 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.468305] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e6a4fa2-fb77-4a45-b8b8-0fa412a5635b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.482025] env[63202]: DEBUG nova.compute.provider_tree [None req-7d0ec686-adfb-458e-8a0e-e648f13a89a8 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1043.542367] env[63202]: DEBUG nova.compute.utils [None req-785be6b2-df40-4037-b7c3-91924a04c311 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Build of instance 0d7fd297-37db-4876-bb91-5bb0b7dc335e aborted: Failed to rebuild volume backed instance. {{(pid=63202) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 1043.544660] env[63202]: ERROR nova.compute.manager [None req-785be6b2-df40-4037-b7c3-91924a04c311 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Setting instance vm_state to ERROR: nova.exception.BuildAbortException: Build of instance 0d7fd297-37db-4876-bb91-5bb0b7dc335e aborted: Failed to rebuild volume backed instance. [ 1043.544660] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Traceback (most recent call last): [ 1043.544660] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 1043.544660] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] self.driver.rebuild(**kwargs) [ 1043.544660] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] File "/opt/stack/nova/nova/virt/driver.py", line 493, in rebuild [ 1043.544660] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] raise NotImplementedError() [ 1043.544660] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] NotImplementedError [ 1043.544660] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] [ 1043.544660] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] During handling of the above exception, another exception occurred: [ 1043.544660] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] [ 1043.544660] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Traceback (most recent call last): [ 1043.544660] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] File "/opt/stack/nova/nova/compute/manager.py", line 3600, in _rebuild_volume_backed_instance [ 1043.544660] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] self._detach_root_volume(context, instance, root_bdm) [ 1043.545128] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] File "/opt/stack/nova/nova/compute/manager.py", line 3579, in _detach_root_volume [ 1043.545128] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] with excutils.save_and_reraise_exception(): [ 1043.545128] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1043.545128] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] self.force_reraise() [ 1043.545128] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1043.545128] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] raise self.value [ 1043.545128] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 1043.545128] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] self.driver.detach_volume(context, old_connection_info, [ 1043.545128] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 559, in detach_volume [ 1043.545128] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] return self._volumeops.detach_volume(connection_info, instance) [ 1043.545128] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 1043.545128] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] self._detach_volume_vmdk(connection_info, instance) [ 1043.545605] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 1043.545605] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 1043.545605] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 1043.545605] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] stable_ref.fetch_moref(session) [ 1043.545605] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 1043.545605] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] raise exception.InstanceNotFound(instance_id=self._uuid) [ 1043.545605] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] nova.exception.InstanceNotFound: Instance 0d7fd297-37db-4876-bb91-5bb0b7dc335e could not be found. [ 1043.545605] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] [ 1043.545605] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] During handling of the above exception, another exception occurred: [ 1043.545605] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] [ 1043.545605] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Traceback (most recent call last): [ 1043.545605] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] File "/opt/stack/nova/nova/compute/manager.py", line 10865, in _error_out_instance_on_exception [ 1043.545605] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] yield [ 1043.545605] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] File "/opt/stack/nova/nova/compute/manager.py", line 3868, in rebuild_instance [ 1043.546106] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] self._do_rebuild_instance_with_claim( [ 1043.546106] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] File "/opt/stack/nova/nova/compute/manager.py", line 3954, in _do_rebuild_instance_with_claim [ 1043.546106] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] self._do_rebuild_instance( [ 1043.546106] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] File "/opt/stack/nova/nova/compute/manager.py", line 4146, in _do_rebuild_instance [ 1043.546106] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] self._rebuild_default_impl(**kwargs) [ 1043.546106] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] File "/opt/stack/nova/nova/compute/manager.py", line 3723, in _rebuild_default_impl [ 1043.546106] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] self._rebuild_volume_backed_instance( [ 1043.546106] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] File "/opt/stack/nova/nova/compute/manager.py", line 3615, in _rebuild_volume_backed_instance [ 1043.546106] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] raise exception.BuildAbortException( [ 1043.546106] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] nova.exception.BuildAbortException: Build of instance 0d7fd297-37db-4876-bb91-5bb0b7dc335e aborted: Failed to rebuild volume backed instance. [ 1043.546106] env[63202]: ERROR nova.compute.manager [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] [ 1043.674289] env[63202]: DEBUG nova.network.neutron [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Successfully created port: b78164a7-4413-4943-b4fd-0a55d84f87fe {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1043.755932] env[63202]: DEBUG oslo_vmware.api [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385824, 'name': Rename_Task, 'duration_secs': 0.156368} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.756296] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1043.756417] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-32367ae3-1cab-43e6-ad53-9260f068db8c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.763151] env[63202]: DEBUG oslo_vmware.api [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 1043.763151] env[63202]: value = "task-1385825" [ 1043.763151] env[63202]: _type = "Task" [ 1043.763151] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.770620] env[63202]: DEBUG oslo_vmware.api [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385825, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.785396] env[63202]: DEBUG nova.compute.manager [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1043.919279] env[63202]: DEBUG oslo_vmware.api [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]524fab16-02de-f3ad-5334-147bf0261254, 'name': SearchDatastore_Task, 'duration_secs': 0.009774} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.920112] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-36cb2771-08bc-4544-8754-b2c49edb0109 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.926326] env[63202]: DEBUG oslo_vmware.api [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Waiting for the task: (returnval){ [ 1043.926326] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]526404db-d63f-11d0-7b69-ceffb23c63bd" [ 1043.926326] env[63202]: _type = "Task" [ 1043.926326] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.934292] env[63202]: DEBUG oslo_vmware.api [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]526404db-d63f-11d0-7b69-ceffb23c63bd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.986935] env[63202]: DEBUG nova.scheduler.client.report [None req-7d0ec686-adfb-458e-8a0e-e648f13a89a8 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1044.274179] env[63202]: DEBUG oslo_vmware.api [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385825, 'name': PowerOnVM_Task, 'duration_secs': 0.413774} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.274439] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1044.274678] env[63202]: INFO nova.compute.manager [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Took 7.58 seconds to spawn the instance on the hypervisor. [ 1044.274858] env[63202]: DEBUG nova.compute.manager [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1044.275683] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-214ac2e0-7886-4b8c-9bf8-7a7b3a22cc63 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.437302] env[63202]: DEBUG oslo_vmware.api [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]526404db-d63f-11d0-7b69-ceffb23c63bd, 'name': SearchDatastore_Task, 'duration_secs': 0.008624} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.437560] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.437827] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] e1e2e8a4-80aa-45eb-a90e-d20cba2943ee/e1e2e8a4-80aa-45eb-a90e-d20cba2943ee.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1044.438125] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7f148b70-4e75-41de-8444-7a1aa52f1cbc {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.445029] env[63202]: DEBUG oslo_vmware.api [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Waiting for the task: (returnval){ [ 1044.445029] env[63202]: value = "task-1385827" [ 1044.445029] env[63202]: _type = "Task" [ 1044.445029] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.453759] env[63202]: DEBUG oslo_vmware.api [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385827, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.492357] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7d0ec686-adfb-458e-8a0e-e648f13a89a8 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.721s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.495036] env[63202]: DEBUG oslo_concurrency.lockutils [None req-43190153-e4ba-44d9-9717-98230b065546 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.023s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.495036] env[63202]: DEBUG nova.objects.instance [None req-43190153-e4ba-44d9-9717-98230b065546 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lazy-loading 'resources' on Instance uuid 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1044.514126] env[63202]: INFO nova.scheduler.client.report [None req-7d0ec686-adfb-458e-8a0e-e648f13a89a8 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Deleted allocations for instance c2b6f3a4-7414-4fc4-893b-9b613fc5e381 [ 1044.793776] env[63202]: DEBUG nova.compute.manager [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1044.796967] env[63202]: INFO nova.compute.manager [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Took 13.84 seconds to build instance. [ 1044.821174] env[63202]: DEBUG nova.virt.hardware [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1044.821470] env[63202]: DEBUG nova.virt.hardware [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1044.821633] env[63202]: DEBUG nova.virt.hardware [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1044.821819] env[63202]: DEBUG nova.virt.hardware [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1044.821982] env[63202]: DEBUG nova.virt.hardware [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1044.822142] env[63202]: DEBUG nova.virt.hardware [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1044.822395] env[63202]: DEBUG nova.virt.hardware [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1044.822576] env[63202]: DEBUG nova.virt.hardware [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1044.822751] env[63202]: DEBUG nova.virt.hardware [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1044.822914] env[63202]: DEBUG nova.virt.hardware [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1044.823110] env[63202]: DEBUG nova.virt.hardware [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1044.824089] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd4242ca-2629-452e-8f23-96ad47bd18eb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.834716] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b812207-4584-412a-ae04-e55f5ab38d20 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.955934] env[63202]: DEBUG oslo_vmware.api [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385827, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.021827] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7d0ec686-adfb-458e-8a0e-e648f13a89a8 tempest-ServerDiskConfigTestJSON-1517182470 tempest-ServerDiskConfigTestJSON-1517182470-project-member] Lock "c2b6f3a4-7414-4fc4-893b-9b613fc5e381" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.961s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.089913] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d93028f8-e0b3-4ad9-a910-ad40692bd030 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.099148] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0820d106-c64f-4e0e-bbd9-54333f145ae0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.136869] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26551a21-0cb6-41d6-b4ac-ea3b0432110e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.145810] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b915f6a-4e03-4923-b61f-2015ea49dce3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.160341] env[63202]: DEBUG nova.compute.provider_tree [None req-43190153-e4ba-44d9-9717-98230b065546 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1045.212184] env[63202]: DEBUG nova.compute.manager [req-c8ce1fe9-8237-45bf-874a-cae8cf23cfd3 req-aee8297f-163e-4dd6-8fb5-b87692a97928 service nova] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Received event network-vif-plugged-b78164a7-4413-4943-b4fd-0a55d84f87fe {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1045.212446] env[63202]: DEBUG oslo_concurrency.lockutils [req-c8ce1fe9-8237-45bf-874a-cae8cf23cfd3 req-aee8297f-163e-4dd6-8fb5-b87692a97928 service nova] Acquiring lock "1123a12e-5218-415d-b286-2f005fe57b29-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.212662] env[63202]: DEBUG oslo_concurrency.lockutils [req-c8ce1fe9-8237-45bf-874a-cae8cf23cfd3 req-aee8297f-163e-4dd6-8fb5-b87692a97928 service nova] Lock "1123a12e-5218-415d-b286-2f005fe57b29-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.212832] env[63202]: DEBUG oslo_concurrency.lockutils [req-c8ce1fe9-8237-45bf-874a-cae8cf23cfd3 req-aee8297f-163e-4dd6-8fb5-b87692a97928 service nova] Lock "1123a12e-5218-415d-b286-2f005fe57b29-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.213013] env[63202]: DEBUG nova.compute.manager [req-c8ce1fe9-8237-45bf-874a-cae8cf23cfd3 req-aee8297f-163e-4dd6-8fb5-b87692a97928 service nova] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] No waiting events found dispatching network-vif-plugged-b78164a7-4413-4943-b4fd-0a55d84f87fe {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1045.213485] env[63202]: WARNING nova.compute.manager [req-c8ce1fe9-8237-45bf-874a-cae8cf23cfd3 req-aee8297f-163e-4dd6-8fb5-b87692a97928 service nova] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Received unexpected event network-vif-plugged-b78164a7-4413-4943-b4fd-0a55d84f87fe for instance with vm_state building and task_state spawning. [ 1045.299150] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a56be8ca-e756-4190-beb8-33f44bf20672 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "3a65ddea-2d3d-45a1-a175-c8fad756793d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.350s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.412394] env[63202]: DEBUG nova.network.neutron [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Successfully updated port: b78164a7-4413-4943-b4fd-0a55d84f87fe {{(pid=63202) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1045.457286] env[63202]: DEBUG oslo_vmware.api [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385827, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.512566} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.457552] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] e1e2e8a4-80aa-45eb-a90e-d20cba2943ee/e1e2e8a4-80aa-45eb-a90e-d20cba2943ee.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1045.458611] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1045.458611] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fe3fe1e9-02b2-4545-89a4-7c32f7fb5eaf {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.465474] env[63202]: DEBUG oslo_vmware.api [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Waiting for the task: (returnval){ [ 1045.465474] env[63202]: value = "task-1385828" [ 1045.465474] env[63202]: _type = "Task" [ 1045.465474] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.474701] env[63202]: DEBUG oslo_vmware.api [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385828, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.560657] env[63202]: DEBUG oslo_concurrency.lockutils [None req-785be6b2-df40-4037-b7c3-91924a04c311 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.663139] env[63202]: DEBUG nova.scheduler.client.report [None req-43190153-e4ba-44d9-9717-98230b065546 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1045.915695] env[63202]: DEBUG oslo_concurrency.lockutils [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Acquiring lock "refresh_cache-1123a12e-5218-415d-b286-2f005fe57b29" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.917377] env[63202]: DEBUG oslo_concurrency.lockutils [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Acquired lock "refresh_cache-1123a12e-5218-415d-b286-2f005fe57b29" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.917377] env[63202]: DEBUG nova.network.neutron [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1045.975978] env[63202]: DEBUG oslo_vmware.api [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385828, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067139} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.976270] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1045.977033] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6af7a465-d43a-4697-9edf-460908f92907 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.000985] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] e1e2e8a4-80aa-45eb-a90e-d20cba2943ee/e1e2e8a4-80aa-45eb-a90e-d20cba2943ee.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1046.001328] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-020f410c-eef2-4bac-b7d2-85fc736fdb3b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.015552] env[63202]: DEBUG oslo_concurrency.lockutils [None req-39947364-c013-4bae-a10c-fe55d8de4aa1 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "3a65ddea-2d3d-45a1-a175-c8fad756793d" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.015796] env[63202]: DEBUG oslo_concurrency.lockutils [None req-39947364-c013-4bae-a10c-fe55d8de4aa1 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "3a65ddea-2d3d-45a1-a175-c8fad756793d" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.015971] env[63202]: DEBUG nova.compute.manager [None req-39947364-c013-4bae-a10c-fe55d8de4aa1 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1046.016797] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-708d810d-bdd4-4636-9533-5bbee386c11c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.023499] env[63202]: DEBUG nova.compute.manager [None req-39947364-c013-4bae-a10c-fe55d8de4aa1 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63202) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1046.024143] env[63202]: DEBUG nova.objects.instance [None req-39947364-c013-4bae-a10c-fe55d8de4aa1 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lazy-loading 'flavor' on Instance uuid 3a65ddea-2d3d-45a1-a175-c8fad756793d {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1046.028746] env[63202]: DEBUG oslo_vmware.api [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Waiting for the task: (returnval){ [ 1046.028746] env[63202]: value = "task-1385829" [ 1046.028746] env[63202]: _type = "Task" [ 1046.028746] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.036653] env[63202]: DEBUG oslo_vmware.api [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385829, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.168343] env[63202]: DEBUG oslo_concurrency.lockutils [None req-43190153-e4ba-44d9-9717-98230b065546 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.674s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.172222] env[63202]: DEBUG oslo_concurrency.lockutils [None req-785be6b2-df40-4037-b7c3-91924a04c311 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.612s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.194226] env[63202]: INFO nova.scheduler.client.report [None req-43190153-e4ba-44d9-9717-98230b065546 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Deleted allocations for instance 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b [ 1046.279348] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a91c7fe-281f-44e0-8fac-862f23a6ccf8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.288822] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ae751c2-ac87-4414-a398-2fc06a2cf840 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.322199] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a46c22ba-bad7-4851-b26f-bc35391befbb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.330056] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94b73546-60e0-44de-9b8f-40a60bbfb439 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.344018] env[63202]: DEBUG nova.compute.provider_tree [None req-785be6b2-df40-4037-b7c3-91924a04c311 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1046.352352] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b45da5e-6ca2-4fbd-a329-bb0b66911ee3 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Acquiring lock "0d7fd297-37db-4876-bb91-5bb0b7dc335e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.352627] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b45da5e-6ca2-4fbd-a329-bb0b66911ee3 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Lock "0d7fd297-37db-4876-bb91-5bb0b7dc335e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.352826] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b45da5e-6ca2-4fbd-a329-bb0b66911ee3 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Acquiring lock "0d7fd297-37db-4876-bb91-5bb0b7dc335e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.353022] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b45da5e-6ca2-4fbd-a329-bb0b66911ee3 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Lock "0d7fd297-37db-4876-bb91-5bb0b7dc335e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.353200] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b45da5e-6ca2-4fbd-a329-bb0b66911ee3 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Lock "0d7fd297-37db-4876-bb91-5bb0b7dc335e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.355300] env[63202]: INFO nova.compute.manager [None req-5b45da5e-6ca2-4fbd-a329-bb0b66911ee3 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Terminating instance [ 1046.357158] env[63202]: DEBUG nova.compute.manager [None req-5b45da5e-6ca2-4fbd-a329-bb0b66911ee3 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1046.357433] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d2d8dcc0-65d2-428a-8f3f-424e38564c81 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.366143] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c722c9c-4463-469d-9a5f-575a6b920fa1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.391837] env[63202]: WARNING nova.virt.vmwareapi.driver [None req-5b45da5e-6ca2-4fbd-a329-bb0b66911ee3 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 0d7fd297-37db-4876-bb91-5bb0b7dc335e could not be found. [ 1046.392098] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-5b45da5e-6ca2-4fbd-a329-bb0b66911ee3 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1046.392417] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-75856dbe-4208-40d9-9497-a278772fd51a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.400339] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61d0a1ab-138c-4f33-bf2f-5b93295e0a67 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.427661] env[63202]: WARNING nova.virt.vmwareapi.vmops [None req-5b45da5e-6ca2-4fbd-a329-bb0b66911ee3 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0d7fd297-37db-4876-bb91-5bb0b7dc335e could not be found. [ 1046.427661] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-5b45da5e-6ca2-4fbd-a329-bb0b66911ee3 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1046.427661] env[63202]: INFO nova.compute.manager [None req-5b45da5e-6ca2-4fbd-a329-bb0b66911ee3 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Took 0.07 seconds to destroy the instance on the hypervisor. [ 1046.428088] env[63202]: DEBUG oslo.service.loopingcall [None req-5b45da5e-6ca2-4fbd-a329-bb0b66911ee3 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1046.428088] env[63202]: DEBUG nova.compute.manager [-] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1046.428158] env[63202]: DEBUG nova.network.neutron [-] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1046.475689] env[63202]: DEBUG nova.network.neutron [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1046.536936] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-39947364-c013-4bae-a10c-fe55d8de4aa1 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1046.537568] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a197d4a9-ca14-42d5-8e89-70b51007f716 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.545611] env[63202]: DEBUG oslo_vmware.api [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385829, 'name': ReconfigVM_Task, 'duration_secs': 0.295893} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.549969] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Reconfigured VM instance instance-00000062 to attach disk [datastore1] e1e2e8a4-80aa-45eb-a90e-d20cba2943ee/e1e2e8a4-80aa-45eb-a90e-d20cba2943ee.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1046.549969] env[63202]: DEBUG oslo_vmware.api [None req-39947364-c013-4bae-a10c-fe55d8de4aa1 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 1046.549969] env[63202]: value = "task-1385830" [ 1046.549969] env[63202]: _type = "Task" [ 1046.549969] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.549969] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c2bb48e2-dabd-468e-a4e1-8d25a9c5beac {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.562796] env[63202]: DEBUG oslo_vmware.api [None req-39947364-c013-4bae-a10c-fe55d8de4aa1 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385830, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.569425] env[63202]: DEBUG oslo_vmware.api [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Waiting for the task: (returnval){ [ 1046.569425] env[63202]: value = "task-1385831" [ 1046.569425] env[63202]: _type = "Task" [ 1046.569425] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.586047] env[63202]: DEBUG oslo_vmware.api [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385831, 'name': Rename_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.708718] env[63202]: DEBUG oslo_concurrency.lockutils [None req-43190153-e4ba-44d9-9717-98230b065546 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.212s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.729197] env[63202]: DEBUG nova.network.neutron [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Updating instance_info_cache with network_info: [{"id": "b78164a7-4413-4943-b4fd-0a55d84f87fe", "address": "fa:16:3e:e4:94:7c", "network": {"id": "d36e22d7-c866-4005-8049-ebce7c0376bd", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1130666168-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "9c1a31c777da417c96b2552ca50dc3a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e614f8e-6b11-4b6b-a421-904bca6acd91", "external-id": "nsx-vlan-transportzone-923", "segmentation_id": 923, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb78164a7-44", "ovs_interfaceid": "b78164a7-4413-4943-b4fd-0a55d84f87fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1046.846948] env[63202]: DEBUG nova.scheduler.client.report [None req-785be6b2-df40-4037-b7c3-91924a04c311 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1047.061352] env[63202]: DEBUG oslo_vmware.api [None req-39947364-c013-4bae-a10c-fe55d8de4aa1 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385830, 'name': PowerOffVM_Task, 'duration_secs': 0.19401} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.061685] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-39947364-c013-4bae-a10c-fe55d8de4aa1 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1047.062597] env[63202]: DEBUG nova.compute.manager [None req-39947364-c013-4bae-a10c-fe55d8de4aa1 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1047.063324] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-566d5912-38b4-48b7-9f4c-be45aeffa5f2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.082861] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7e8bf74e-2f29-4e8e-833c-300c690c58d3 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquiring lock "3c6d286d-c152-46f1-b212-96a67324a56d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.083804] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7e8bf74e-2f29-4e8e-833c-300c690c58d3 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "3c6d286d-c152-46f1-b212-96a67324a56d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.084010] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7e8bf74e-2f29-4e8e-833c-300c690c58d3 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquiring lock "3c6d286d-c152-46f1-b212-96a67324a56d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.085609] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7e8bf74e-2f29-4e8e-833c-300c690c58d3 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "3c6d286d-c152-46f1-b212-96a67324a56d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.085609] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7e8bf74e-2f29-4e8e-833c-300c690c58d3 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "3c6d286d-c152-46f1-b212-96a67324a56d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.087206] env[63202]: DEBUG oslo_vmware.api [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385831, 'name': Rename_Task, 'duration_secs': 0.22501} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.091021] env[63202]: INFO nova.compute.manager [None req-7e8bf74e-2f29-4e8e-833c-300c690c58d3 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Terminating instance [ 1047.091021] env[63202]: DEBUG nova.compute.manager [None req-7e8bf74e-2f29-4e8e-833c-300c690c58d3 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1047.091021] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-7e8bf74e-2f29-4e8e-833c-300c690c58d3 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1047.091021] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1047.091458] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-624c65a2-6392-49dc-9865-856d98b63ef3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.095353] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c95c7fc3-bb49-4f64-944a-dbba7c64099b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.102575] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e8bf74e-2f29-4e8e-833c-300c690c58d3 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1047.104299] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9cc44f9d-41a6-45ca-80c7-4a705f37e57d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.106347] env[63202]: DEBUG oslo_vmware.api [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Waiting for the task: (returnval){ [ 1047.106347] env[63202]: value = "task-1385832" [ 1047.106347] env[63202]: _type = "Task" [ 1047.106347] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.112704] env[63202]: DEBUG oslo_vmware.api [None req-7e8bf74e-2f29-4e8e-833c-300c690c58d3 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 1047.112704] env[63202]: value = "task-1385833" [ 1047.112704] env[63202]: _type = "Task" [ 1047.112704] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.119990] env[63202]: DEBUG oslo_vmware.api [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385832, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.131568] env[63202]: DEBUG oslo_vmware.api [None req-7e8bf74e-2f29-4e8e-833c-300c690c58d3 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385833, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.231844] env[63202]: DEBUG oslo_concurrency.lockutils [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Releasing lock "refresh_cache-1123a12e-5218-415d-b286-2f005fe57b29" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1047.232256] env[63202]: DEBUG nova.compute.manager [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Instance network_info: |[{"id": "b78164a7-4413-4943-b4fd-0a55d84f87fe", "address": "fa:16:3e:e4:94:7c", "network": {"id": "d36e22d7-c866-4005-8049-ebce7c0376bd", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1130666168-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "9c1a31c777da417c96b2552ca50dc3a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e614f8e-6b11-4b6b-a421-904bca6acd91", "external-id": "nsx-vlan-transportzone-923", "segmentation_id": 923, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb78164a7-44", "ovs_interfaceid": "b78164a7-4413-4943-b4fd-0a55d84f87fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1047.232697] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e4:94:7c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2e614f8e-6b11-4b6b-a421-904bca6acd91', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b78164a7-4413-4943-b4fd-0a55d84f87fe', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1047.243140] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Creating folder: Project (9c1a31c777da417c96b2552ca50dc3a7). Parent ref: group-v294090. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1047.243944] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0aae7022-4342-402f-8831-1e7a4d43582f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.256314] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Created folder: Project (9c1a31c777da417c96b2552ca50dc3a7) in parent group-v294090. [ 1047.257446] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Creating folder: Instances. Parent ref: group-v294212. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1047.257724] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6d0355b3-53db-4b5d-9c62-9cf530183f11 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.269425] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Created folder: Instances in parent group-v294212. [ 1047.270091] env[63202]: DEBUG oslo.service.loopingcall [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1047.270091] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1047.270371] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-457257a2-40de-4fb9-b84f-07d05024a4cb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.297120] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1047.297120] env[63202]: value = "task-1385836" [ 1047.297120] env[63202]: _type = "Task" [ 1047.297120] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.304530] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385836, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.309022] env[63202]: DEBUG nova.compute.manager [req-11429a6e-bb41-45f1-92a9-a6d2b249e796 req-538fc45e-1f5c-4ac5-8223-c6a9f31a3a99 service nova] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Received event network-vif-deleted-119b092b-6a1b-41ad-aa1c-b06e28b28f75 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1047.309022] env[63202]: INFO nova.compute.manager [req-11429a6e-bb41-45f1-92a9-a6d2b249e796 req-538fc45e-1f5c-4ac5-8223-c6a9f31a3a99 service nova] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Neutron deleted interface 119b092b-6a1b-41ad-aa1c-b06e28b28f75; detaching it from the instance and deleting it from the info cache [ 1047.309022] env[63202]: DEBUG nova.network.neutron [req-11429a6e-bb41-45f1-92a9-a6d2b249e796 req-538fc45e-1f5c-4ac5-8223-c6a9f31a3a99 service nova] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1047.324746] env[63202]: DEBUG nova.compute.manager [req-ba5f4095-bc90-4696-bf48-ecd85548ec59 req-2f83b28e-260e-4183-bcce-ede267cde8da service nova] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Received event network-changed-b78164a7-4413-4943-b4fd-0a55d84f87fe {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1047.325113] env[63202]: DEBUG nova.compute.manager [req-ba5f4095-bc90-4696-bf48-ecd85548ec59 req-2f83b28e-260e-4183-bcce-ede267cde8da service nova] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Refreshing instance network info cache due to event network-changed-b78164a7-4413-4943-b4fd-0a55d84f87fe. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1047.325565] env[63202]: DEBUG oslo_concurrency.lockutils [req-ba5f4095-bc90-4696-bf48-ecd85548ec59 req-2f83b28e-260e-4183-bcce-ede267cde8da service nova] Acquiring lock "refresh_cache-1123a12e-5218-415d-b286-2f005fe57b29" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.325715] env[63202]: DEBUG oslo_concurrency.lockutils [req-ba5f4095-bc90-4696-bf48-ecd85548ec59 req-2f83b28e-260e-4183-bcce-ede267cde8da service nova] Acquired lock "refresh_cache-1123a12e-5218-415d-b286-2f005fe57b29" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.325877] env[63202]: DEBUG nova.network.neutron [req-ba5f4095-bc90-4696-bf48-ecd85548ec59 req-2f83b28e-260e-4183-bcce-ede267cde8da service nova] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Refreshing network info cache for port b78164a7-4413-4943-b4fd-0a55d84f87fe {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1047.352893] env[63202]: DEBUG oslo_concurrency.lockutils [None req-785be6b2-df40-4037-b7c3-91924a04c311 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.181s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.353187] env[63202]: INFO nova.compute.manager [None req-785be6b2-df40-4037-b7c3-91924a04c311 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Successfully reverted task state from rebuilding on failure for instance. [ 1047.403323] env[63202]: DEBUG oslo_concurrency.lockutils [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.403577] env[63202]: DEBUG oslo_concurrency.lockutils [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.403791] env[63202]: DEBUG nova.objects.instance [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lazy-loading 'pci_requests' on Instance uuid 8437f856-a707-49c5-b8eb-5a22cdb990f8 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1047.577170] env[63202]: DEBUG oslo_concurrency.lockutils [None req-39947364-c013-4bae-a10c-fe55d8de4aa1 tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "3a65ddea-2d3d-45a1-a175-c8fad756793d" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.561s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.618724] env[63202]: DEBUG oslo_vmware.api [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385832, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.629079] env[63202]: DEBUG oslo_vmware.api [None req-7e8bf74e-2f29-4e8e-833c-300c690c58d3 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385833, 'name': PowerOffVM_Task, 'duration_secs': 0.201188} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.633795] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e8bf74e-2f29-4e8e-833c-300c690c58d3 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1047.633995] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-7e8bf74e-2f29-4e8e-833c-300c690c58d3 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1047.634294] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4672ee1e-c193-4290-b4f2-4ebf3f61739f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.660057] env[63202]: DEBUG nova.network.neutron [-] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1047.695000] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-7e8bf74e-2f29-4e8e-833c-300c690c58d3 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1047.695233] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-7e8bf74e-2f29-4e8e-833c-300c690c58d3 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Deleting contents of the VM from datastore datastore2 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1047.695424] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e8bf74e-2f29-4e8e-833c-300c690c58d3 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Deleting the datastore file [datastore2] 3c6d286d-c152-46f1-b212-96a67324a56d {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1047.695715] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fab5a76c-71b5-42e3-8d6b-49d38605ed2d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.704582] env[63202]: DEBUG oslo_vmware.api [None req-7e8bf74e-2f29-4e8e-833c-300c690c58d3 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 1047.704582] env[63202]: value = "task-1385838" [ 1047.704582] env[63202]: _type = "Task" [ 1047.704582] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.713737] env[63202]: DEBUG oslo_vmware.api [None req-7e8bf74e-2f29-4e8e-833c-300c690c58d3 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385838, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.807694] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385836, 'name': CreateVM_Task, 'duration_secs': 0.337183} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.808066] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1047.809163] env[63202]: DEBUG oslo_concurrency.lockutils [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.809485] env[63202]: DEBUG oslo_concurrency.lockutils [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.810101] env[63202]: DEBUG oslo_concurrency.lockutils [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1047.810889] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-54e52721-a418-4c64-a2c7-2dc20402f253 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.814022] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c33274e0-7e57-44fd-8271-bb8afef53e9d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.819371] env[63202]: DEBUG oslo_vmware.api [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Waiting for the task: (returnval){ [ 1047.819371] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52e20857-f1db-f008-905c-2011f9c0a861" [ 1047.819371] env[63202]: _type = "Task" [ 1047.819371] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.831831] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5fd1701-86e0-4001-b7fe-162b9dce667a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.854174] env[63202]: DEBUG oslo_vmware.api [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52e20857-f1db-f008-905c-2011f9c0a861, 'name': SearchDatastore_Task, 'duration_secs': 0.011991} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.854497] env[63202]: DEBUG oslo_concurrency.lockutils [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1047.854731] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1047.854971] env[63202]: DEBUG oslo_concurrency.lockutils [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.855143] env[63202]: DEBUG oslo_concurrency.lockutils [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.855326] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1047.855808] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ece6c17f-4f06-4775-8340-a58e3ac17056 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.864573] env[63202]: DEBUG nova.compute.manager [req-11429a6e-bb41-45f1-92a9-a6d2b249e796 req-538fc45e-1f5c-4ac5-8223-c6a9f31a3a99 service nova] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Detach interface failed, port_id=119b092b-6a1b-41ad-aa1c-b06e28b28f75, reason: Instance 0d7fd297-37db-4876-bb91-5bb0b7dc335e could not be found. {{(pid=63202) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1047.870806] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1047.870987] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1047.871744] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1dfb8865-a82b-4168-9c64-b730f7651a0d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.876824] env[63202]: DEBUG oslo_vmware.api [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Waiting for the task: (returnval){ [ 1047.876824] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]5226fd20-fe21-fd61-c412-c07ff6696b89" [ 1047.876824] env[63202]: _type = "Task" [ 1047.876824] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.884442] env[63202]: DEBUG oslo_vmware.api [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]5226fd20-fe21-fd61-c412-c07ff6696b89, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.907685] env[63202]: DEBUG nova.objects.instance [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lazy-loading 'numa_topology' on Instance uuid 8437f856-a707-49c5-b8eb-5a22cdb990f8 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1048.014756] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "439ee94c-04f2-45d3-a486-81a216c8db4e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.014916] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "439ee94c-04f2-45d3-a486-81a216c8db4e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.118616] env[63202]: DEBUG oslo_vmware.api [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385832, 'name': PowerOnVM_Task, 'duration_secs': 0.529462} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.118938] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1048.119212] env[63202]: INFO nova.compute.manager [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Took 9.17 seconds to spawn the instance on the hypervisor. [ 1048.119440] env[63202]: DEBUG nova.compute.manager [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1048.120279] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3790fd0a-efbf-4cd2-abfc-0f4f074ea651 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.164102] env[63202]: INFO nova.compute.manager [-] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Took 1.73 seconds to deallocate network for instance. [ 1048.174395] env[63202]: DEBUG nova.network.neutron [req-ba5f4095-bc90-4696-bf48-ecd85548ec59 req-2f83b28e-260e-4183-bcce-ede267cde8da service nova] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Updated VIF entry in instance network info cache for port b78164a7-4413-4943-b4fd-0a55d84f87fe. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1048.174395] env[63202]: DEBUG nova.network.neutron [req-ba5f4095-bc90-4696-bf48-ecd85548ec59 req-2f83b28e-260e-4183-bcce-ede267cde8da service nova] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Updating instance_info_cache with network_info: [{"id": "b78164a7-4413-4943-b4fd-0a55d84f87fe", "address": "fa:16:3e:e4:94:7c", "network": {"id": "d36e22d7-c866-4005-8049-ebce7c0376bd", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1130666168-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "9c1a31c777da417c96b2552ca50dc3a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e614f8e-6b11-4b6b-a421-904bca6acd91", "external-id": "nsx-vlan-transportzone-923", "segmentation_id": 923, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb78164a7-44", "ovs_interfaceid": "b78164a7-4413-4943-b4fd-0a55d84f87fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1048.216878] env[63202]: DEBUG oslo_vmware.api [None req-7e8bf74e-2f29-4e8e-833c-300c690c58d3 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385838, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14444} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.217157] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e8bf74e-2f29-4e8e-833c-300c690c58d3 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1048.217407] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-7e8bf74e-2f29-4e8e-833c-300c690c58d3 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Deleted contents of the VM from datastore datastore2 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1048.217498] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-7e8bf74e-2f29-4e8e-833c-300c690c58d3 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1048.217668] env[63202]: INFO nova.compute.manager [None req-7e8bf74e-2f29-4e8e-833c-300c690c58d3 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1048.218585] env[63202]: DEBUG oslo.service.loopingcall [None req-7e8bf74e-2f29-4e8e-833c-300c690c58d3 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1048.218585] env[63202]: DEBUG nova.compute.manager [-] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1048.218585] env[63202]: DEBUG nova.network.neutron [-] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1048.389583] env[63202]: DEBUG oslo_vmware.api [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]5226fd20-fe21-fd61-c412-c07ff6696b89, 'name': SearchDatastore_Task, 'duration_secs': 0.008476} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.389791] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9dab9f4f-7c78-4678-9175-9d5b987a22ee {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.397266] env[63202]: DEBUG oslo_vmware.api [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Waiting for the task: (returnval){ [ 1048.397266] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]528d141e-8a58-17d5-35df-96d29c654341" [ 1048.397266] env[63202]: _type = "Task" [ 1048.397266] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.412740] env[63202]: INFO nova.compute.claims [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1048.415601] env[63202]: DEBUG oslo_vmware.api [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]528d141e-8a58-17d5-35df-96d29c654341, 'name': SearchDatastore_Task, 'duration_secs': 0.009352} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.416391] env[63202]: DEBUG oslo_concurrency.lockutils [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1048.416687] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] 1123a12e-5218-415d-b286-2f005fe57b29/1123a12e-5218-415d-b286-2f005fe57b29.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1048.417314] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4c497516-a773-4686-adbb-aa32b11d176a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.427090] env[63202]: DEBUG oslo_vmware.api [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Waiting for the task: (returnval){ [ 1048.427090] env[63202]: value = "task-1385839" [ 1048.427090] env[63202]: _type = "Task" [ 1048.427090] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.436149] env[63202]: DEBUG oslo_vmware.api [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Task: {'id': task-1385839, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.517427] env[63202]: DEBUG nova.compute.manager [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1048.641149] env[63202]: INFO nova.compute.manager [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Took 14.78 seconds to build instance. [ 1048.676610] env[63202]: DEBUG oslo_concurrency.lockutils [req-ba5f4095-bc90-4696-bf48-ecd85548ec59 req-2f83b28e-260e-4183-bcce-ede267cde8da service nova] Releasing lock "refresh_cache-1123a12e-5218-415d-b286-2f005fe57b29" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1048.749432] env[63202]: INFO nova.compute.manager [None req-5b45da5e-6ca2-4fbd-a329-bb0b66911ee3 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Took 0.59 seconds to detach 1 volumes for instance. [ 1048.752021] env[63202]: DEBUG nova.compute.manager [None req-5b45da5e-6ca2-4fbd-a329-bb0b66911ee3 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Deleting volume: 6518fe7c-9ed9-49fb-b419-6c72d538c63d {{(pid=63202) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 1048.906716] env[63202]: DEBUG oslo_concurrency.lockutils [None req-705cb5ce-f8c8-46a2-9fa9-d43a39abb16b tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "3a65ddea-2d3d-45a1-a175-c8fad756793d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.907132] env[63202]: DEBUG oslo_concurrency.lockutils [None req-705cb5ce-f8c8-46a2-9fa9-d43a39abb16b tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "3a65ddea-2d3d-45a1-a175-c8fad756793d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.907242] env[63202]: DEBUG oslo_concurrency.lockutils [None req-705cb5ce-f8c8-46a2-9fa9-d43a39abb16b tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "3a65ddea-2d3d-45a1-a175-c8fad756793d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.907394] env[63202]: DEBUG oslo_concurrency.lockutils [None req-705cb5ce-f8c8-46a2-9fa9-d43a39abb16b tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "3a65ddea-2d3d-45a1-a175-c8fad756793d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.907562] env[63202]: DEBUG oslo_concurrency.lockutils [None req-705cb5ce-f8c8-46a2-9fa9-d43a39abb16b tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "3a65ddea-2d3d-45a1-a175-c8fad756793d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.913031] env[63202]: INFO nova.compute.manager [None req-705cb5ce-f8c8-46a2-9fa9-d43a39abb16b tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Terminating instance [ 1048.914033] env[63202]: DEBUG nova.compute.manager [None req-705cb5ce-f8c8-46a2-9fa9-d43a39abb16b tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1048.914228] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-705cb5ce-f8c8-46a2-9fa9-d43a39abb16b tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1048.915091] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31816050-46de-488e-873d-83a40552bba5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.925891] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-705cb5ce-f8c8-46a2-9fa9-d43a39abb16b tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1048.926307] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-58821ecc-8de8-478e-bf06-ab701eec8ae5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.938350] env[63202]: DEBUG oslo_vmware.api [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Task: {'id': task-1385839, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.428443} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.938591] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] 1123a12e-5218-415d-b286-2f005fe57b29/1123a12e-5218-415d-b286-2f005fe57b29.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1048.938793] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1048.939032] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7f263d49-58f5-4eea-ad37-82df6de79cef {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.945793] env[63202]: DEBUG oslo_vmware.api [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Waiting for the task: (returnval){ [ 1048.945793] env[63202]: value = "task-1385842" [ 1048.945793] env[63202]: _type = "Task" [ 1048.945793] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.953518] env[63202]: DEBUG oslo_vmware.api [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Task: {'id': task-1385842, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.017987] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-705cb5ce-f8c8-46a2-9fa9-d43a39abb16b tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1049.017987] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-705cb5ce-f8c8-46a2-9fa9-d43a39abb16b tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Deleting contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1049.017987] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-705cb5ce-f8c8-46a2-9fa9-d43a39abb16b tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Deleting the datastore file [datastore1] 3a65ddea-2d3d-45a1-a175-c8fad756793d {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1049.017987] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d414f11b-387c-4862-adcc-19837d5440da {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.025707] env[63202]: DEBUG oslo_vmware.api [None req-705cb5ce-f8c8-46a2-9fa9-d43a39abb16b tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for the task: (returnval){ [ 1049.025707] env[63202]: value = "task-1385843" [ 1049.025707] env[63202]: _type = "Task" [ 1049.025707] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.038657] env[63202]: DEBUG oslo_vmware.api [None req-705cb5ce-f8c8-46a2-9fa9-d43a39abb16b tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385843, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.051426] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.144122] env[63202]: DEBUG oslo_concurrency.lockutils [None req-bd4187dd-e40e-411e-afeb-1e4eb0475787 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lock "e1e2e8a4-80aa-45eb-a90e-d20cba2943ee" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.297s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.298162] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b45da5e-6ca2-4fbd-a329-bb0b66911ee3 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.333988] env[63202]: DEBUG nova.compute.manager [req-f1cfe8e6-39f2-4229-bf03-680e38708de4 req-dfa758d6-7873-4e66-a469-2dc22f7ab509 service nova] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Received event network-vif-deleted-131d0949-eb5d-4ee6-a244-d7898f8e7b0f {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1049.334210] env[63202]: INFO nova.compute.manager [req-f1cfe8e6-39f2-4229-bf03-680e38708de4 req-dfa758d6-7873-4e66-a469-2dc22f7ab509 service nova] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Neutron deleted interface 131d0949-eb5d-4ee6-a244-d7898f8e7b0f; detaching it from the instance and deleting it from the info cache [ 1049.334447] env[63202]: DEBUG nova.network.neutron [req-f1cfe8e6-39f2-4229-bf03-680e38708de4 req-dfa758d6-7873-4e66-a469-2dc22f7ab509 service nova] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1049.339558] env[63202]: DEBUG nova.network.neutron [-] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1049.457884] env[63202]: DEBUG oslo_vmware.api [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Task: {'id': task-1385842, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.056839} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.458441] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1049.459355] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd585a6c-cab0-4680-86b6-33591031a15c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.491808] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] 1123a12e-5218-415d-b286-2f005fe57b29/1123a12e-5218-415d-b286-2f005fe57b29.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1049.492482] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1739ce0a-0df6-4450-b55b-c95fa99f29f8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.519039] env[63202]: DEBUG oslo_vmware.api [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Waiting for the task: (returnval){ [ 1049.519039] env[63202]: value = "task-1385844" [ 1049.519039] env[63202]: _type = "Task" [ 1049.519039] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.528557] env[63202]: DEBUG oslo_vmware.api [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Task: {'id': task-1385844, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.536815] env[63202]: DEBUG oslo_vmware.api [None req-705cb5ce-f8c8-46a2-9fa9-d43a39abb16b tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Task: {'id': task-1385843, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.173073} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.542049] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-705cb5ce-f8c8-46a2-9fa9-d43a39abb16b tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1049.542049] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-705cb5ce-f8c8-46a2-9fa9-d43a39abb16b tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Deleted contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1049.542049] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-705cb5ce-f8c8-46a2-9fa9-d43a39abb16b tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1049.542049] env[63202]: INFO nova.compute.manager [None req-705cb5ce-f8c8-46a2-9fa9-d43a39abb16b tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Took 0.63 seconds to destroy the instance on the hypervisor. [ 1049.542049] env[63202]: DEBUG oslo.service.loopingcall [None req-705cb5ce-f8c8-46a2-9fa9-d43a39abb16b tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1049.544349] env[63202]: DEBUG nova.compute.manager [-] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1049.544349] env[63202]: DEBUG nova.network.neutron [-] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1049.592040] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73e95924-e032-4bf2-9158-f994a0356b32 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.600398] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a809744-6b9b-45a5-91dd-9e48c515c18a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.640022] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c07a8c78-fc57-4402-ba2e-58bb6df30953 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.647512] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60f2a5ab-c4f0-4c3d-8ab4-360e1f9296d6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.664152] env[63202]: DEBUG nova.compute.provider_tree [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1049.842023] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0fcdaa95-3c35-4193-a3b2-0025e2cfcada {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.842023] env[63202]: INFO nova.compute.manager [-] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Took 1.62 seconds to deallocate network for instance. [ 1049.850533] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-542478cd-ca00-4ad6-a66e-1a6fabc6a3c8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.877923] env[63202]: DEBUG nova.compute.manager [req-f1cfe8e6-39f2-4229-bf03-680e38708de4 req-dfa758d6-7873-4e66-a469-2dc22f7ab509 service nova] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Detach interface failed, port_id=131d0949-eb5d-4ee6-a244-d7898f8e7b0f, reason: Instance 3c6d286d-c152-46f1-b212-96a67324a56d could not be found. {{(pid=63202) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1050.029102] env[63202]: DEBUG oslo_vmware.api [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Task: {'id': task-1385844, 'name': ReconfigVM_Task, 'duration_secs': 0.268458} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.029778] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Reconfigured VM instance instance-00000063 to attach disk [datastore1] 1123a12e-5218-415d-b286-2f005fe57b29/1123a12e-5218-415d-b286-2f005fe57b29.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1050.030612] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ecd4e51b-d61c-4513-8b8b-117aac65951a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.039492] env[63202]: DEBUG oslo_vmware.api [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Waiting for the task: (returnval){ [ 1050.039492] env[63202]: value = "task-1385845" [ 1050.039492] env[63202]: _type = "Task" [ 1050.039492] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.046794] env[63202]: DEBUG oslo_vmware.api [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Task: {'id': task-1385845, 'name': Rename_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.169142] env[63202]: DEBUG nova.scheduler.client.report [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1050.349935] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7e8bf74e-2f29-4e8e-833c-300c690c58d3 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1050.547736] env[63202]: DEBUG oslo_vmware.api [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Task: {'id': task-1385845, 'name': Rename_Task, 'duration_secs': 0.191934} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.548364] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1050.548364] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4e720d41-fea1-4246-bd80-e39f8d297dd3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.555650] env[63202]: DEBUG oslo_vmware.api [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Waiting for the task: (returnval){ [ 1050.555650] env[63202]: value = "task-1385846" [ 1050.555650] env[63202]: _type = "Task" [ 1050.555650] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.564145] env[63202]: DEBUG oslo_vmware.api [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Task: {'id': task-1385846, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.673331] env[63202]: DEBUG oslo_concurrency.lockutils [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.269s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.678121] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.628s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.680066] env[63202]: INFO nova.compute.claims [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1050.734995] env[63202]: INFO nova.network.neutron [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Updating port 07cc827a-876a-4432-a5fc-9ba8920dc5f9 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1050.776273] env[63202]: DEBUG nova.network.neutron [-] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1051.065547] env[63202]: DEBUG oslo_vmware.api [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Task: {'id': task-1385846, 'name': PowerOnVM_Task, 'duration_secs': 0.487834} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.065839] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1051.066054] env[63202]: INFO nova.compute.manager [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Took 6.27 seconds to spawn the instance on the hypervisor. [ 1051.066243] env[63202]: DEBUG nova.compute.manager [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1051.067042] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea2ac3c5-48d9-4827-ab3b-1121ca4eaacb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.284095] env[63202]: INFO nova.compute.manager [-] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Took 1.74 seconds to deallocate network for instance. [ 1051.369382] env[63202]: DEBUG nova.compute.manager [req-44805614-1f55-4595-ade4-44d237e7b066 req-1faf2352-16ba-4c6e-9432-508e80e92ae2 service nova] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Received event network-vif-deleted-3bfc4703-af3e-41fe-a35a-a5a8fadf1aef {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1051.369712] env[63202]: DEBUG nova.compute.manager [req-44805614-1f55-4595-ade4-44d237e7b066 req-1faf2352-16ba-4c6e-9432-508e80e92ae2 service nova] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Received event network-changed-3cec3f36-0fb8-45c0-be6e-48cd6cf58aa3 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1051.369956] env[63202]: DEBUG nova.compute.manager [req-44805614-1f55-4595-ade4-44d237e7b066 req-1faf2352-16ba-4c6e-9432-508e80e92ae2 service nova] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Refreshing instance network info cache due to event network-changed-3cec3f36-0fb8-45c0-be6e-48cd6cf58aa3. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1051.370867] env[63202]: DEBUG oslo_concurrency.lockutils [req-44805614-1f55-4595-ade4-44d237e7b066 req-1faf2352-16ba-4c6e-9432-508e80e92ae2 service nova] Acquiring lock "refresh_cache-e1e2e8a4-80aa-45eb-a90e-d20cba2943ee" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1051.371180] env[63202]: DEBUG oslo_concurrency.lockutils [req-44805614-1f55-4595-ade4-44d237e7b066 req-1faf2352-16ba-4c6e-9432-508e80e92ae2 service nova] Acquired lock "refresh_cache-e1e2e8a4-80aa-45eb-a90e-d20cba2943ee" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.371364] env[63202]: DEBUG nova.network.neutron [req-44805614-1f55-4595-ade4-44d237e7b066 req-1faf2352-16ba-4c6e-9432-508e80e92ae2 service nova] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Refreshing network info cache for port 3cec3f36-0fb8-45c0-be6e-48cd6cf58aa3 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1051.584723] env[63202]: INFO nova.compute.manager [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Took 11.54 seconds to build instance. [ 1051.792738] env[63202]: DEBUG oslo_concurrency.lockutils [None req-705cb5ce-f8c8-46a2-9fa9-d43a39abb16b tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.820052] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b245d2f1-d6f8-4bd3-87fa-06fdf3da8715 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.827833] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-674e1edf-b97b-4685-9e32-29bb0b3004aa {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.859766] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-688bfe44-9cfb-4119-975d-8b3ce20e9033 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.868406] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c64b69b-0369-46fd-988b-b6b279a2a5b4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.882480] env[63202]: DEBUG nova.compute.provider_tree [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1052.086802] env[63202]: DEBUG oslo_concurrency.lockutils [None req-aa183cd1-9991-43be-ba9d-981e650febeb tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Lock "1123a12e-5218-415d-b286-2f005fe57b29" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.051s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.211293] env[63202]: DEBUG nova.network.neutron [req-44805614-1f55-4595-ade4-44d237e7b066 req-1faf2352-16ba-4c6e-9432-508e80e92ae2 service nova] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Updated VIF entry in instance network info cache for port 3cec3f36-0fb8-45c0-be6e-48cd6cf58aa3. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1052.211964] env[63202]: DEBUG nova.network.neutron [req-44805614-1f55-4595-ade4-44d237e7b066 req-1faf2352-16ba-4c6e-9432-508e80e92ae2 service nova] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Updating instance_info_cache with network_info: [{"id": "3cec3f36-0fb8-45c0-be6e-48cd6cf58aa3", "address": "fa:16:3e:ac:21:28", "network": {"id": "345538f4-e791-4fc6-9719-237c11f1382a", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-449651008-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.223", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a9e939cdf4a94a8aa0107a5761771c2c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa09e855-8af1-419b-b78d-8ffcc94b1bfb", "external-id": "nsx-vlan-transportzone-901", "segmentation_id": 901, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cec3f36-0f", "ovs_interfaceid": "3cec3f36-0fb8-45c0-be6e-48cd6cf58aa3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1052.391173] env[63202]: DEBUG nova.scheduler.client.report [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1052.458060] env[63202]: DEBUG oslo_concurrency.lockutils [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquiring lock "refresh_cache-8437f856-a707-49c5-b8eb-5a22cdb990f8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1052.458237] env[63202]: DEBUG oslo_concurrency.lockutils [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquired lock "refresh_cache-8437f856-a707-49c5-b8eb-5a22cdb990f8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1052.458407] env[63202]: DEBUG nova.network.neutron [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1052.612883] env[63202]: INFO nova.compute.manager [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Rescuing [ 1052.613854] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Acquiring lock "refresh_cache-1123a12e-5218-415d-b286-2f005fe57b29" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1052.613968] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Acquired lock "refresh_cache-1123a12e-5218-415d-b286-2f005fe57b29" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1052.614133] env[63202]: DEBUG nova.network.neutron [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1052.715986] env[63202]: DEBUG oslo_concurrency.lockutils [req-44805614-1f55-4595-ade4-44d237e7b066 req-1faf2352-16ba-4c6e-9432-508e80e92ae2 service nova] Releasing lock "refresh_cache-e1e2e8a4-80aa-45eb-a90e-d20cba2943ee" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1052.895773] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.218s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.896326] env[63202]: DEBUG nova.compute.manager [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1052.898841] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b45da5e-6ca2-4fbd-a329-bb0b66911ee3 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.601s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.899073] env[63202]: DEBUG nova.objects.instance [None req-5b45da5e-6ca2-4fbd-a329-bb0b66911ee3 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Lazy-loading 'resources' on Instance uuid 0d7fd297-37db-4876-bb91-5bb0b7dc335e {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1053.172467] env[63202]: DEBUG nova.network.neutron [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Updating instance_info_cache with network_info: [{"id": "07cc827a-876a-4432-a5fc-9ba8920dc5f9", "address": "fa:16:3e:71:e6:9f", "network": {"id": "06ab5813-9ad9-4021-9bdb-f2f02af8d73f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1714653503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.175", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b2de35030a9484094e964ffc30a822d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3093647a-bab7-4562-ada0-428725e8c0fc", "external-id": "nsx-vlan-transportzone-660", "segmentation_id": 660, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap07cc827a-87", "ovs_interfaceid": "07cc827a-876a-4432-a5fc-9ba8920dc5f9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1053.203078] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Acquiring lock "6e172a2a-1918-430c-84ce-005fef97bbf1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.203322] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Lock "6e172a2a-1918-430c-84ce-005fef97bbf1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.312988] env[63202]: DEBUG nova.network.neutron [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Updating instance_info_cache with network_info: [{"id": "b78164a7-4413-4943-b4fd-0a55d84f87fe", "address": "fa:16:3e:e4:94:7c", "network": {"id": "d36e22d7-c866-4005-8049-ebce7c0376bd", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1130666168-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "9c1a31c777da417c96b2552ca50dc3a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e614f8e-6b11-4b6b-a421-904bca6acd91", "external-id": "nsx-vlan-transportzone-923", "segmentation_id": 923, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb78164a7-44", "ovs_interfaceid": "b78164a7-4413-4943-b4fd-0a55d84f87fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1053.401205] env[63202]: DEBUG nova.compute.manager [req-16e6ad8b-2333-4b6b-bcc6-88c0676fbfaf req-30a1576c-df0b-4402-978b-5f916e3d84d6 service nova] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Received event network-vif-plugged-07cc827a-876a-4432-a5fc-9ba8920dc5f9 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1053.401580] env[63202]: DEBUG oslo_concurrency.lockutils [req-16e6ad8b-2333-4b6b-bcc6-88c0676fbfaf req-30a1576c-df0b-4402-978b-5f916e3d84d6 service nova] Acquiring lock "8437f856-a707-49c5-b8eb-5a22cdb990f8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.401782] env[63202]: DEBUG oslo_concurrency.lockutils [req-16e6ad8b-2333-4b6b-bcc6-88c0676fbfaf req-30a1576c-df0b-4402-978b-5f916e3d84d6 service nova] Lock "8437f856-a707-49c5-b8eb-5a22cdb990f8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.401847] env[63202]: DEBUG oslo_concurrency.lockutils [req-16e6ad8b-2333-4b6b-bcc6-88c0676fbfaf req-30a1576c-df0b-4402-978b-5f916e3d84d6 service nova] Lock "8437f856-a707-49c5-b8eb-5a22cdb990f8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.402030] env[63202]: DEBUG nova.compute.manager [req-16e6ad8b-2333-4b6b-bcc6-88c0676fbfaf req-30a1576c-df0b-4402-978b-5f916e3d84d6 service nova] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] No waiting events found dispatching network-vif-plugged-07cc827a-876a-4432-a5fc-9ba8920dc5f9 {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1053.402205] env[63202]: WARNING nova.compute.manager [req-16e6ad8b-2333-4b6b-bcc6-88c0676fbfaf req-30a1576c-df0b-4402-978b-5f916e3d84d6 service nova] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Received unexpected event network-vif-plugged-07cc827a-876a-4432-a5fc-9ba8920dc5f9 for instance with vm_state shelved_offloaded and task_state spawning. [ 1053.402385] env[63202]: DEBUG nova.compute.manager [req-16e6ad8b-2333-4b6b-bcc6-88c0676fbfaf req-30a1576c-df0b-4402-978b-5f916e3d84d6 service nova] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Received event network-changed-07cc827a-876a-4432-a5fc-9ba8920dc5f9 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1053.402559] env[63202]: DEBUG nova.compute.manager [req-16e6ad8b-2333-4b6b-bcc6-88c0676fbfaf req-30a1576c-df0b-4402-978b-5f916e3d84d6 service nova] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Refreshing instance network info cache due to event network-changed-07cc827a-876a-4432-a5fc-9ba8920dc5f9. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1053.402727] env[63202]: DEBUG oslo_concurrency.lockutils [req-16e6ad8b-2333-4b6b-bcc6-88c0676fbfaf req-30a1576c-df0b-4402-978b-5f916e3d84d6 service nova] Acquiring lock "refresh_cache-8437f856-a707-49c5-b8eb-5a22cdb990f8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1053.403777] env[63202]: DEBUG nova.compute.utils [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1053.407980] env[63202]: DEBUG nova.compute.manager [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1053.408154] env[63202]: DEBUG nova.network.neutron [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1053.450697] env[63202]: DEBUG nova.policy [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b59e7e62d92e4f5eb0fee41e8f5196d7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f508ec04edc844a19640a8a85f27e5b9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 1053.526426] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15f6f998-91fb-487d-9b94-143de219d0c9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.534289] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f207246f-5688-452d-bb65-c14da5a15c76 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.565682] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cd450cf-736c-4b25-bb88-53ff1624c78f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.573277] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f4fb4ed-8006-4cbd-9be6-d09d7bb0d1f9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.586761] env[63202]: DEBUG nova.compute.provider_tree [None req-5b45da5e-6ca2-4fbd-a329-bb0b66911ee3 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1053.674301] env[63202]: DEBUG oslo_concurrency.lockutils [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Releasing lock "refresh_cache-8437f856-a707-49c5-b8eb-5a22cdb990f8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1053.676895] env[63202]: DEBUG oslo_concurrency.lockutils [req-16e6ad8b-2333-4b6b-bcc6-88c0676fbfaf req-30a1576c-df0b-4402-978b-5f916e3d84d6 service nova] Acquired lock "refresh_cache-8437f856-a707-49c5-b8eb-5a22cdb990f8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1053.677126] env[63202]: DEBUG nova.network.neutron [req-16e6ad8b-2333-4b6b-bcc6-88c0676fbfaf req-30a1576c-df0b-4402-978b-5f916e3d84d6 service nova] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Refreshing network info cache for port 07cc827a-876a-4432-a5fc-9ba8920dc5f9 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1053.705649] env[63202]: DEBUG nova.compute.manager [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1053.713452] env[63202]: DEBUG nova.virt.hardware [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='b7c3940c24a6b8d66f44381838c17bee',container_format='bare',created_at=2024-10-31T10:30:02Z,direct_url=,disk_format='vmdk',id=3db9d5e2-204c-47e7-ae45-3ed10283296d,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-2119583611-shelved',owner='2b2de35030a9484094e964ffc30a822d',properties=ImageMetaProps,protected=,size=31669248,status='active',tags=,updated_at=2024-10-31T10:30:15Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1053.713452] env[63202]: DEBUG nova.virt.hardware [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1053.713452] env[63202]: DEBUG nova.virt.hardware [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1053.713685] env[63202]: DEBUG nova.virt.hardware [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1053.713685] env[63202]: DEBUG nova.virt.hardware [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1053.713685] env[63202]: DEBUG nova.virt.hardware [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1053.714470] env[63202]: DEBUG nova.virt.hardware [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1053.714704] env[63202]: DEBUG nova.virt.hardware [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1053.714888] env[63202]: DEBUG nova.virt.hardware [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1053.715068] env[63202]: DEBUG nova.virt.hardware [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1053.715333] env[63202]: DEBUG nova.virt.hardware [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1053.716100] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-668bac00-dedf-4c24-8f84-85f3128aad14 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.727194] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7444f05-382e-4e54-b702-6a5bb77a6e9c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.743665] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:71:e6:9f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3093647a-bab7-4562-ada0-428725e8c0fc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '07cc827a-876a-4432-a5fc-9ba8920dc5f9', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1053.751488] env[63202]: DEBUG oslo.service.loopingcall [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1053.751779] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1053.752014] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fe4b0ec9-727a-4818-8e82-b4b870fb213f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.773121] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1053.773121] env[63202]: value = "task-1385847" [ 1053.773121] env[63202]: _type = "Task" [ 1053.773121] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.776704] env[63202]: DEBUG nova.network.neutron [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Successfully created port: a8eb92d1-368f-42cc-83d2-d516202801a3 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1053.783395] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385847, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.816312] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Releasing lock "refresh_cache-1123a12e-5218-415d-b286-2f005fe57b29" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1053.910342] env[63202]: DEBUG nova.compute.manager [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1054.090559] env[63202]: DEBUG nova.scheduler.client.report [None req-5b45da5e-6ca2-4fbd-a329-bb0b66911ee3 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1054.224799] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.286980] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385847, 'name': CreateVM_Task, 'duration_secs': 0.376933} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.287287] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1054.288164] env[63202]: DEBUG oslo_concurrency.lockutils [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3db9d5e2-204c-47e7-ae45-3ed10283296d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1054.288438] env[63202]: DEBUG oslo_concurrency.lockutils [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3db9d5e2-204c-47e7-ae45-3ed10283296d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1054.288916] env[63202]: DEBUG oslo_concurrency.lockutils [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3db9d5e2-204c-47e7-ae45-3ed10283296d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1054.289256] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7e794a16-9327-4dde-a91f-c399341bf96c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.294484] env[63202]: DEBUG oslo_vmware.api [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1054.294484] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52413d25-d841-61b3-4c2f-3bc31539b5a1" [ 1054.294484] env[63202]: _type = "Task" [ 1054.294484] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.302238] env[63202]: DEBUG oslo_vmware.api [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52413d25-d841-61b3-4c2f-3bc31539b5a1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.363055] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1054.363329] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bc822f0b-3968-42e6-9b73-7ffe1501e6d1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.370271] env[63202]: DEBUG oslo_vmware.api [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Waiting for the task: (returnval){ [ 1054.370271] env[63202]: value = "task-1385848" [ 1054.370271] env[63202]: _type = "Task" [ 1054.370271] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.380190] env[63202]: DEBUG oslo_vmware.api [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Task: {'id': task-1385848, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.442539] env[63202]: DEBUG nova.network.neutron [req-16e6ad8b-2333-4b6b-bcc6-88c0676fbfaf req-30a1576c-df0b-4402-978b-5f916e3d84d6 service nova] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Updated VIF entry in instance network info cache for port 07cc827a-876a-4432-a5fc-9ba8920dc5f9. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1054.443370] env[63202]: DEBUG nova.network.neutron [req-16e6ad8b-2333-4b6b-bcc6-88c0676fbfaf req-30a1576c-df0b-4402-978b-5f916e3d84d6 service nova] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Updating instance_info_cache with network_info: [{"id": "07cc827a-876a-4432-a5fc-9ba8920dc5f9", "address": "fa:16:3e:71:e6:9f", "network": {"id": "06ab5813-9ad9-4021-9bdb-f2f02af8d73f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1714653503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.175", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b2de35030a9484094e964ffc30a822d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3093647a-bab7-4562-ada0-428725e8c0fc", "external-id": "nsx-vlan-transportzone-660", "segmentation_id": 660, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap07cc827a-87", "ovs_interfaceid": "07cc827a-876a-4432-a5fc-9ba8920dc5f9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1054.596059] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b45da5e-6ca2-4fbd-a329-bb0b66911ee3 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.697s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.599053] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7e8bf74e-2f29-4e8e-833c-300c690c58d3 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.249s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.599318] env[63202]: DEBUG nova.objects.instance [None req-7e8bf74e-2f29-4e8e-833c-300c690c58d3 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lazy-loading 'resources' on Instance uuid 3c6d286d-c152-46f1-b212-96a67324a56d {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1054.804427] env[63202]: DEBUG oslo_concurrency.lockutils [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3db9d5e2-204c-47e7-ae45-3ed10283296d" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1054.804693] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Processing image 3db9d5e2-204c-47e7-ae45-3ed10283296d {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1054.804935] env[63202]: DEBUG oslo_concurrency.lockutils [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3db9d5e2-204c-47e7-ae45-3ed10283296d/3db9d5e2-204c-47e7-ae45-3ed10283296d.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1054.805100] env[63202]: DEBUG oslo_concurrency.lockutils [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3db9d5e2-204c-47e7-ae45-3ed10283296d/3db9d5e2-204c-47e7-ae45-3ed10283296d.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1054.805282] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1054.805531] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-97196fd8-516f-4502-b416-9c966f95b051 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.813817] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1054.813992] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1054.814696] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c444122-c7be-41fc-9ac8-b3f4190c3087 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.819670] env[63202]: DEBUG oslo_vmware.api [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1054.819670] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52a1a1f3-d70d-9237-f315-e1e644311070" [ 1054.819670] env[63202]: _type = "Task" [ 1054.819670] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.827072] env[63202]: DEBUG oslo_vmware.api [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52a1a1f3-d70d-9237-f315-e1e644311070, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.880022] env[63202]: DEBUG oslo_vmware.api [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Task: {'id': task-1385848, 'name': PowerOffVM_Task, 'duration_secs': 0.213805} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.880292] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1054.881071] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dac7744-a6ed-4b84-8693-3ffff43a135c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.898617] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf023a45-ac6c-4d5c-8f90-d043fa62ccbf {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.926144] env[63202]: DEBUG nova.compute.manager [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1054.929586] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1054.929996] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f9307fde-43b0-48c8-9142-c5ccf984c7e8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.936577] env[63202]: DEBUG oslo_vmware.api [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Waiting for the task: (returnval){ [ 1054.936577] env[63202]: value = "task-1385849" [ 1054.936577] env[63202]: _type = "Task" [ 1054.936577] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.947690] env[63202]: DEBUG oslo_concurrency.lockutils [req-16e6ad8b-2333-4b6b-bcc6-88c0676fbfaf req-30a1576c-df0b-4402-978b-5f916e3d84d6 service nova] Releasing lock "refresh_cache-8437f856-a707-49c5-b8eb-5a22cdb990f8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1054.949797] env[63202]: DEBUG nova.virt.hardware [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1054.950025] env[63202]: DEBUG nova.virt.hardware [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1054.950190] env[63202]: DEBUG nova.virt.hardware [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1054.950376] env[63202]: DEBUG nova.virt.hardware [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1054.950527] env[63202]: DEBUG nova.virt.hardware [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1054.950678] env[63202]: DEBUG nova.virt.hardware [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1054.950882] env[63202]: DEBUG nova.virt.hardware [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1054.951053] env[63202]: DEBUG nova.virt.hardware [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1054.951225] env[63202]: DEBUG nova.virt.hardware [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1054.951417] env[63202]: DEBUG nova.virt.hardware [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1054.951596] env[63202]: DEBUG nova.virt.hardware [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1054.951894] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] VM already powered off {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1054.952082] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1054.952335] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1054.952523] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1054.952709] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1054.953501] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fa6731c-35b1-4524-9baa-37cafc2537bd {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.955762] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5c57ec1f-71c5-4890-b532-bc169159142e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.962492] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d22f9119-6f5f-42f7-9d8e-6d442e669311 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.967212] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1054.967395] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1054.968389] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b931425-4b4a-420f-917d-1a49799c392e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.980918] env[63202]: DEBUG oslo_vmware.api [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Waiting for the task: (returnval){ [ 1054.980918] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]5206400d-9d3d-c27f-2399-37596314f825" [ 1054.980918] env[63202]: _type = "Task" [ 1054.980918] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.987880] env[63202]: DEBUG oslo_vmware.api [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]5206400d-9d3d-c27f-2399-37596314f825, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.121913] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5b45da5e-6ca2-4fbd-a329-bb0b66911ee3 tempest-ServerActionsV293TestJSON-1100877798 tempest-ServerActionsV293TestJSON-1100877798-project-member] Lock "0d7fd297-37db-4876-bb91-5bb0b7dc335e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.769s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.228880] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afc22929-f878-4bf1-b890-6fdc65920ac4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.236769] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83e47420-a3b8-4fcc-a2f0-04a6f327f64b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.268227] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f4b4277-ca03-4f4a-88e6-8c5647dfae6a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.276960] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb3cab75-524d-410f-8c8d-89b3ba685be1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.291808] env[63202]: DEBUG nova.compute.provider_tree [None req-7e8bf74e-2f29-4e8e-833c-300c690c58d3 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1055.296342] env[63202]: DEBUG nova.compute.manager [req-f559cdd7-a3ba-4a07-8f32-86c236268d78 req-50c4f37f-b2c2-425b-b379-9129bbf88dd6 service nova] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Received event network-vif-plugged-a8eb92d1-368f-42cc-83d2-d516202801a3 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1055.296342] env[63202]: DEBUG oslo_concurrency.lockutils [req-f559cdd7-a3ba-4a07-8f32-86c236268d78 req-50c4f37f-b2c2-425b-b379-9129bbf88dd6 service nova] Acquiring lock "439ee94c-04f2-45d3-a486-81a216c8db4e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.296482] env[63202]: DEBUG oslo_concurrency.lockutils [req-f559cdd7-a3ba-4a07-8f32-86c236268d78 req-50c4f37f-b2c2-425b-b379-9129bbf88dd6 service nova] Lock "439ee94c-04f2-45d3-a486-81a216c8db4e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.296646] env[63202]: DEBUG oslo_concurrency.lockutils [req-f559cdd7-a3ba-4a07-8f32-86c236268d78 req-50c4f37f-b2c2-425b-b379-9129bbf88dd6 service nova] Lock "439ee94c-04f2-45d3-a486-81a216c8db4e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.296864] env[63202]: DEBUG nova.compute.manager [req-f559cdd7-a3ba-4a07-8f32-86c236268d78 req-50c4f37f-b2c2-425b-b379-9129bbf88dd6 service nova] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] No waiting events found dispatching network-vif-plugged-a8eb92d1-368f-42cc-83d2-d516202801a3 {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1055.297057] env[63202]: WARNING nova.compute.manager [req-f559cdd7-a3ba-4a07-8f32-86c236268d78 req-50c4f37f-b2c2-425b-b379-9129bbf88dd6 service nova] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Received unexpected event network-vif-plugged-a8eb92d1-368f-42cc-83d2-d516202801a3 for instance with vm_state building and task_state spawning. [ 1055.331210] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Preparing fetch location {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1055.331485] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Fetch image to [datastore1] OSTACK_IMG_0b4c512d-eb92-4472-b3c8-e79a01cd4464/OSTACK_IMG_0b4c512d-eb92-4472-b3c8-e79a01cd4464.vmdk {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1055.331673] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Downloading stream optimized image 3db9d5e2-204c-47e7-ae45-3ed10283296d to [datastore1] OSTACK_IMG_0b4c512d-eb92-4472-b3c8-e79a01cd4464/OSTACK_IMG_0b4c512d-eb92-4472-b3c8-e79a01cd4464.vmdk on the data store datastore1 as vApp {{(pid=63202) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1055.331840] env[63202]: DEBUG nova.virt.vmwareapi.images [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Downloading image file data 3db9d5e2-204c-47e7-ae45-3ed10283296d to the ESX as VM named 'OSTACK_IMG_0b4c512d-eb92-4472-b3c8-e79a01cd4464' {{(pid=63202) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1055.385818] env[63202]: DEBUG nova.network.neutron [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Successfully updated port: a8eb92d1-368f-42cc-83d2-d516202801a3 {{(pid=63202) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1055.409371] env[63202]: DEBUG oslo_vmware.rw_handles [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1055.409371] env[63202]: value = "resgroup-9" [ 1055.409371] env[63202]: _type = "ResourcePool" [ 1055.409371] env[63202]: }. {{(pid=63202) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1055.409664] env[63202]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-3c8d6df2-aaf5-4a7d-a505-b7fb30161243 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.430309] env[63202]: DEBUG oslo_vmware.rw_handles [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lease: (returnval){ [ 1055.430309] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52183ab7-31dc-b919-b891-5cd13e3b47ae" [ 1055.430309] env[63202]: _type = "HttpNfcLease" [ 1055.430309] env[63202]: } obtained for vApp import into resource pool (val){ [ 1055.430309] env[63202]: value = "resgroup-9" [ 1055.430309] env[63202]: _type = "ResourcePool" [ 1055.430309] env[63202]: }. {{(pid=63202) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1055.430715] env[63202]: DEBUG oslo_vmware.api [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the lease: (returnval){ [ 1055.430715] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52183ab7-31dc-b919-b891-5cd13e3b47ae" [ 1055.430715] env[63202]: _type = "HttpNfcLease" [ 1055.430715] env[63202]: } to be ready. {{(pid=63202) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1055.436996] env[63202]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1055.436996] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52183ab7-31dc-b919-b891-5cd13e3b47ae" [ 1055.436996] env[63202]: _type = "HttpNfcLease" [ 1055.436996] env[63202]: } is initializing. {{(pid=63202) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1055.491976] env[63202]: DEBUG oslo_vmware.api [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]5206400d-9d3d-c27f-2399-37596314f825, 'name': SearchDatastore_Task, 'duration_secs': 0.008979} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.492953] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a5d93bb1-6509-4ce9-a0e5-16332715dea9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.498014] env[63202]: DEBUG oslo_vmware.api [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Waiting for the task: (returnval){ [ 1055.498014] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52509c8f-5859-62b8-9abf-0e513c8b6814" [ 1055.498014] env[63202]: _type = "Task" [ 1055.498014] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.505724] env[63202]: DEBUG oslo_vmware.api [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52509c8f-5859-62b8-9abf-0e513c8b6814, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.796042] env[63202]: DEBUG nova.scheduler.client.report [None req-7e8bf74e-2f29-4e8e-833c-300c690c58d3 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1055.888220] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "refresh_cache-439ee94c-04f2-45d3-a486-81a216c8db4e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1055.888704] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquired lock "refresh_cache-439ee94c-04f2-45d3-a486-81a216c8db4e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1055.888704] env[63202]: DEBUG nova.network.neutron [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1055.939328] env[63202]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1055.939328] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52183ab7-31dc-b919-b891-5cd13e3b47ae" [ 1055.939328] env[63202]: _type = "HttpNfcLease" [ 1055.939328] env[63202]: } is initializing. {{(pid=63202) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1056.008623] env[63202]: DEBUG oslo_vmware.api [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52509c8f-5859-62b8-9abf-0e513c8b6814, 'name': SearchDatastore_Task, 'duration_secs': 0.008766} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.008958] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1056.009270] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] 1123a12e-5218-415d-b286-2f005fe57b29/bb172aa4-a5a8-4395-9793-2416c30721cd-rescue.vmdk. {{(pid=63202) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1056.009560] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e4a55889-2659-4db4-8683-edde1f50d4b9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.016688] env[63202]: DEBUG oslo_vmware.api [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Waiting for the task: (returnval){ [ 1056.016688] env[63202]: value = "task-1385851" [ 1056.016688] env[63202]: _type = "Task" [ 1056.016688] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.024337] env[63202]: DEBUG oslo_vmware.api [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Task: {'id': task-1385851, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.301151] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7e8bf74e-2f29-4e8e-833c-300c690c58d3 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.702s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.303752] env[63202]: DEBUG oslo_concurrency.lockutils [None req-705cb5ce-f8c8-46a2-9fa9-d43a39abb16b tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.511s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.303986] env[63202]: DEBUG nova.objects.instance [None req-705cb5ce-f8c8-46a2-9fa9-d43a39abb16b tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lazy-loading 'resources' on Instance uuid 3a65ddea-2d3d-45a1-a175-c8fad756793d {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1056.327025] env[63202]: INFO nova.scheduler.client.report [None req-7e8bf74e-2f29-4e8e-833c-300c690c58d3 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Deleted allocations for instance 3c6d286d-c152-46f1-b212-96a67324a56d [ 1056.433609] env[63202]: DEBUG nova.network.neutron [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1056.445225] env[63202]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1056.445225] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52183ab7-31dc-b919-b891-5cd13e3b47ae" [ 1056.445225] env[63202]: _type = "HttpNfcLease" [ 1056.445225] env[63202]: } is ready. {{(pid=63202) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1056.445631] env[63202]: DEBUG oslo_vmware.rw_handles [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1056.445631] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52183ab7-31dc-b919-b891-5cd13e3b47ae" [ 1056.445631] env[63202]: _type = "HttpNfcLease" [ 1056.445631] env[63202]: }. {{(pid=63202) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1056.446435] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86116f07-5531-4072-be1c-72dce9ba5d51 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.455583] env[63202]: DEBUG oslo_vmware.rw_handles [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5256c858-e9a7-517f-225f-a9807ada02fc/disk-0.vmdk from lease info. {{(pid=63202) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1056.455801] env[63202]: DEBUG oslo_vmware.rw_handles [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Creating HTTP connection to write to file with size = 31669248 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5256c858-e9a7-517f-225f-a9807ada02fc/disk-0.vmdk. {{(pid=63202) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1056.526078] env[63202]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-5321d9cb-9a30-43bd-ad7f-fc1918ce6257 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.534712] env[63202]: DEBUG oslo_vmware.api [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Task: {'id': task-1385851, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.427209} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.535658] env[63202]: INFO nova.virt.vmwareapi.ds_util [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] 1123a12e-5218-415d-b286-2f005fe57b29/bb172aa4-a5a8-4395-9793-2416c30721cd-rescue.vmdk. [ 1056.538141] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce0a5526-6a24-4540-94ef-743c309c3b1b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.566486] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] 1123a12e-5218-415d-b286-2f005fe57b29/bb172aa4-a5a8-4395-9793-2416c30721cd-rescue.vmdk or device None with type thin {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1056.569375] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-35ba141c-e0b1-4967-9f81-a5009e6f54db {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.588623] env[63202]: DEBUG oslo_vmware.api [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Waiting for the task: (returnval){ [ 1056.588623] env[63202]: value = "task-1385852" [ 1056.588623] env[63202]: _type = "Task" [ 1056.588623] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.596277] env[63202]: DEBUG oslo_vmware.api [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Task: {'id': task-1385852, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.686793] env[63202]: DEBUG nova.network.neutron [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Updating instance_info_cache with network_info: [{"id": "a8eb92d1-368f-42cc-83d2-d516202801a3", "address": "fa:16:3e:ca:ab:5b", "network": {"id": "5a3b4c9b-2ca7-4f8b-8bbb-ea000db91402", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-543677750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f508ec04edc844a19640a8a85f27e5b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8eb92d1-36", "ovs_interfaceid": "a8eb92d1-368f-42cc-83d2-d516202801a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1056.835350] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7e8bf74e-2f29-4e8e-833c-300c690c58d3 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "3c6d286d-c152-46f1-b212-96a67324a56d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.752s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.920139] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-547d99d4-5863-4fb0-82e5-1be5361d244e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.931540] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c6c25e4-b03b-4f51-a417-85e074bab9e1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.968064] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-517d7282-bab3-4aeb-b536-1ffb39f985e2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.978194] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2524f6c5-414f-4f6c-a85c-548d4c866f60 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.992865] env[63202]: DEBUG nova.compute.provider_tree [None req-705cb5ce-f8c8-46a2-9fa9-d43a39abb16b tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1057.097716] env[63202]: DEBUG oslo_vmware.api [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Task: {'id': task-1385852, 'name': ReconfigVM_Task, 'duration_secs': 0.309693} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.098018] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Reconfigured VM instance instance-00000063 to attach disk [datastore1] 1123a12e-5218-415d-b286-2f005fe57b29/bb172aa4-a5a8-4395-9793-2416c30721cd-rescue.vmdk or device None with type thin {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1057.098842] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59bce3e0-59eb-415a-a721-475eaf54fab1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.126822] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-73c5fdbd-2da0-48e5-b676-c9bfbaa4db99 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.142249] env[63202]: DEBUG oslo_vmware.api [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Waiting for the task: (returnval){ [ 1057.142249] env[63202]: value = "task-1385853" [ 1057.142249] env[63202]: _type = "Task" [ 1057.142249] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.151026] env[63202]: DEBUG oslo_vmware.api [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Task: {'id': task-1385853, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.188804] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Releasing lock "refresh_cache-439ee94c-04f2-45d3-a486-81a216c8db4e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1057.189047] env[63202]: DEBUG nova.compute.manager [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Instance network_info: |[{"id": "a8eb92d1-368f-42cc-83d2-d516202801a3", "address": "fa:16:3e:ca:ab:5b", "network": {"id": "5a3b4c9b-2ca7-4f8b-8bbb-ea000db91402", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-543677750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f508ec04edc844a19640a8a85f27e5b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8eb92d1-36", "ovs_interfaceid": "a8eb92d1-368f-42cc-83d2-d516202801a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1057.189550] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ca:ab:5b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a8eb92d1-368f-42cc-83d2-d516202801a3', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1057.198513] env[63202]: DEBUG oslo.service.loopingcall [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1057.203208] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1057.203502] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0c9f7e31-1268-4437-8439-5043d2bb1937 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.228587] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1057.228587] env[63202]: value = "task-1385854" [ 1057.228587] env[63202]: _type = "Task" [ 1057.228587] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.239737] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385854, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.324151] env[63202]: DEBUG nova.compute.manager [req-979910f7-e6a3-4030-9579-441bb6649a01 req-9867ff8f-a0b3-47a3-9cc7-8e1e12613b2d service nova] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Received event network-changed-a8eb92d1-368f-42cc-83d2-d516202801a3 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1057.324364] env[63202]: DEBUG nova.compute.manager [req-979910f7-e6a3-4030-9579-441bb6649a01 req-9867ff8f-a0b3-47a3-9cc7-8e1e12613b2d service nova] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Refreshing instance network info cache due to event network-changed-a8eb92d1-368f-42cc-83d2-d516202801a3. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1057.324593] env[63202]: DEBUG oslo_concurrency.lockutils [req-979910f7-e6a3-4030-9579-441bb6649a01 req-9867ff8f-a0b3-47a3-9cc7-8e1e12613b2d service nova] Acquiring lock "refresh_cache-439ee94c-04f2-45d3-a486-81a216c8db4e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1057.324791] env[63202]: DEBUG oslo_concurrency.lockutils [req-979910f7-e6a3-4030-9579-441bb6649a01 req-9867ff8f-a0b3-47a3-9cc7-8e1e12613b2d service nova] Acquired lock "refresh_cache-439ee94c-04f2-45d3-a486-81a216c8db4e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.324921] env[63202]: DEBUG nova.network.neutron [req-979910f7-e6a3-4030-9579-441bb6649a01 req-9867ff8f-a0b3-47a3-9cc7-8e1e12613b2d service nova] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Refreshing network info cache for port a8eb92d1-368f-42cc-83d2-d516202801a3 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1057.496551] env[63202]: DEBUG nova.scheduler.client.report [None req-705cb5ce-f8c8-46a2-9fa9-d43a39abb16b tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1057.626820] env[63202]: DEBUG oslo_concurrency.lockutils [None req-707e40ae-6044-442d-9a89-69e1127b3e4d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquiring lock "7437595c-fa35-483e-95f3-b75405b6bd13" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1057.627489] env[63202]: DEBUG oslo_concurrency.lockutils [None req-707e40ae-6044-442d-9a89-69e1127b3e4d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "7437595c-fa35-483e-95f3-b75405b6bd13" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.654035] env[63202]: DEBUG oslo_vmware.api [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Task: {'id': task-1385853, 'name': ReconfigVM_Task, 'duration_secs': 0.213003} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.655497] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1057.655802] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2c6ec72b-449a-4ecd-9060-8262ffb5f376 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.663691] env[63202]: DEBUG oslo_vmware.api [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Waiting for the task: (returnval){ [ 1057.663691] env[63202]: value = "task-1385855" [ 1057.663691] env[63202]: _type = "Task" [ 1057.663691] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.675277] env[63202]: DEBUG oslo_vmware.api [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Task: {'id': task-1385855, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.738325] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385854, 'name': CreateVM_Task, 'duration_secs': 0.465255} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.740265] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1057.741010] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1057.741200] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.741571] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1057.741826] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f743588b-3677-44cf-811c-e173e994695b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.746713] env[63202]: DEBUG oslo_vmware.api [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 1057.746713] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52ea8e07-78c2-1b60-d039-c1f210b23b7a" [ 1057.746713] env[63202]: _type = "Task" [ 1057.746713] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.756365] env[63202]: DEBUG oslo_vmware.api [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52ea8e07-78c2-1b60-d039-c1f210b23b7a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.916951] env[63202]: DEBUG oslo_vmware.rw_handles [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Completed reading data from the image iterator. {{(pid=63202) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1057.917331] env[63202]: DEBUG oslo_vmware.rw_handles [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5256c858-e9a7-517f-225f-a9807ada02fc/disk-0.vmdk. {{(pid=63202) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1057.918448] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-694cf4c7-5e79-45bd-ada2-18e616de2ef0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.925297] env[63202]: DEBUG oslo_vmware.rw_handles [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5256c858-e9a7-517f-225f-a9807ada02fc/disk-0.vmdk is in state: ready. {{(pid=63202) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1057.925549] env[63202]: DEBUG oslo_vmware.rw_handles [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5256c858-e9a7-517f-225f-a9807ada02fc/disk-0.vmdk. {{(pid=63202) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1057.925839] env[63202]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-9caf3ea7-f10b-4420-8031-299746d232b7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.004704] env[63202]: DEBUG oslo_concurrency.lockutils [None req-705cb5ce-f8c8-46a2-9fa9-d43a39abb16b tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.701s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.006730] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.782s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.008313] env[63202]: INFO nova.compute.claims [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1058.027552] env[63202]: INFO nova.scheduler.client.report [None req-705cb5ce-f8c8-46a2-9fa9-d43a39abb16b tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Deleted allocations for instance 3a65ddea-2d3d-45a1-a175-c8fad756793d [ 1058.087618] env[63202]: DEBUG nova.network.neutron [req-979910f7-e6a3-4030-9579-441bb6649a01 req-9867ff8f-a0b3-47a3-9cc7-8e1e12613b2d service nova] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Updated VIF entry in instance network info cache for port a8eb92d1-368f-42cc-83d2-d516202801a3. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1058.087618] env[63202]: DEBUG nova.network.neutron [req-979910f7-e6a3-4030-9579-441bb6649a01 req-9867ff8f-a0b3-47a3-9cc7-8e1e12613b2d service nova] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Updating instance_info_cache with network_info: [{"id": "a8eb92d1-368f-42cc-83d2-d516202801a3", "address": "fa:16:3e:ca:ab:5b", "network": {"id": "5a3b4c9b-2ca7-4f8b-8bbb-ea000db91402", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-543677750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f508ec04edc844a19640a8a85f27e5b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8eb92d1-36", "ovs_interfaceid": "a8eb92d1-368f-42cc-83d2-d516202801a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.132077] env[63202]: INFO nova.compute.manager [None req-707e40ae-6044-442d-9a89-69e1127b3e4d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Detaching volume d296fff5-0fd4-4820-8a58-c19b5b2d0da9 [ 1058.174852] env[63202]: DEBUG oslo_vmware.api [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Task: {'id': task-1385855, 'name': PowerOnVM_Task, 'duration_secs': 0.454183} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.175149] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1058.178114] env[63202]: DEBUG nova.compute.manager [None req-e59424a5-c5f1-44e2-b5a9-a27381a61629 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1058.178915] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b179d7b3-a563-4b53-9f8b-83c48d5c25cb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.182233] env[63202]: INFO nova.virt.block_device [None req-707e40ae-6044-442d-9a89-69e1127b3e4d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Attempting to driver detach volume d296fff5-0fd4-4820-8a58-c19b5b2d0da9 from mountpoint /dev/sdb [ 1058.182462] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-707e40ae-6044-442d-9a89-69e1127b3e4d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Volume detach. Driver type: vmdk {{(pid=63202) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1058.182666] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-707e40ae-6044-442d-9a89-69e1127b3e4d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294189', 'volume_id': 'd296fff5-0fd4-4820-8a58-c19b5b2d0da9', 'name': 'volume-d296fff5-0fd4-4820-8a58-c19b5b2d0da9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7437595c-fa35-483e-95f3-b75405b6bd13', 'attached_at': '', 'detached_at': '', 'volume_id': 'd296fff5-0fd4-4820-8a58-c19b5b2d0da9', 'serial': 'd296fff5-0fd4-4820-8a58-c19b5b2d0da9'} {{(pid=63202) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1058.183771] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6f68363-2486-4316-bb03-76bfbdb9d626 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.208617] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c923e974-868f-43f4-ab16-a8c771d4a6dd {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.215940] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7289a801-b7b2-429e-b482-759f99addb1f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.237050] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9421fc21-ba26-4005-a2a9-23f4dab409c4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.254286] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-707e40ae-6044-442d-9a89-69e1127b3e4d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] The volume has not been displaced from its original location: [datastore1] volume-d296fff5-0fd4-4820-8a58-c19b5b2d0da9/volume-d296fff5-0fd4-4820-8a58-c19b5b2d0da9.vmdk. No consolidation needed. {{(pid=63202) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1058.259959] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-707e40ae-6044-442d-9a89-69e1127b3e4d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Reconfiguring VM instance instance-00000051 to detach disk 2001 {{(pid=63202) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1058.263731] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0b24c38e-53b6-4863-b511-a7e8c0ba7943 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.284133] env[63202]: DEBUG oslo_vmware.api [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52ea8e07-78c2-1b60-d039-c1f210b23b7a, 'name': SearchDatastore_Task, 'duration_secs': 0.016218} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.286400] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1058.286681] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1058.287014] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.287175] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.287300] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1058.287653] env[63202]: DEBUG oslo_vmware.api [None req-707e40ae-6044-442d-9a89-69e1127b3e4d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 1058.287653] env[63202]: value = "task-1385856" [ 1058.287653] env[63202]: _type = "Task" [ 1058.287653] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.288109] env[63202]: DEBUG oslo_vmware.rw_handles [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5256c858-e9a7-517f-225f-a9807ada02fc/disk-0.vmdk. {{(pid=63202) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1058.288383] env[63202]: INFO nova.virt.vmwareapi.images [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Downloaded image file data 3db9d5e2-204c-47e7-ae45-3ed10283296d [ 1058.288535] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cd4f2890-ab81-41a3-89be-d9cb189238d2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.291064] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd28e0c1-440c-4f81-b576-c51814069940 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.316813] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-45c47c10-28e7-478c-afbd-ecb78a91d523 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.318472] env[63202]: DEBUG oslo_vmware.api [None req-707e40ae-6044-442d-9a89-69e1127b3e4d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385856, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.318769] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1058.318912] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1058.319631] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67017c4f-3358-4a44-ac18-f81e42a69ed7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.324488] env[63202]: DEBUG oslo_vmware.api [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 1058.324488] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52ae6816-51c4-c55a-a47b-15d21b660691" [ 1058.324488] env[63202]: _type = "Task" [ 1058.324488] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.332155] env[63202]: DEBUG oslo_vmware.api [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52ae6816-51c4-c55a-a47b-15d21b660691, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.353371] env[63202]: INFO nova.virt.vmwareapi.images [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] The imported VM was unregistered [ 1058.355808] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Caching image {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1058.356075] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Creating directory with path [datastore1] devstack-image-cache_base/3db9d5e2-204c-47e7-ae45-3ed10283296d {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1058.356361] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ecb80f2d-b08e-4dfd-8cda-98b458f01245 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.380481] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Created directory with path [datastore1] devstack-image-cache_base/3db9d5e2-204c-47e7-ae45-3ed10283296d {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1058.380764] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_0b4c512d-eb92-4472-b3c8-e79a01cd4464/OSTACK_IMG_0b4c512d-eb92-4472-b3c8-e79a01cd4464.vmdk to [datastore1] devstack-image-cache_base/3db9d5e2-204c-47e7-ae45-3ed10283296d/3db9d5e2-204c-47e7-ae45-3ed10283296d.vmdk. {{(pid=63202) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1058.381131] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-0f7511c8-ffac-4585-9049-33bd94a3750a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.388754] env[63202]: DEBUG oslo_vmware.api [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1058.388754] env[63202]: value = "task-1385858" [ 1058.388754] env[63202]: _type = "Task" [ 1058.388754] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.396738] env[63202]: DEBUG oslo_vmware.api [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385858, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.536391] env[63202]: DEBUG oslo_concurrency.lockutils [None req-705cb5ce-f8c8-46a2-9fa9-d43a39abb16b tempest-ServersTestJSON-1895684200 tempest-ServersTestJSON-1895684200-project-member] Lock "3a65ddea-2d3d-45a1-a175-c8fad756793d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.629s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.590424] env[63202]: DEBUG oslo_concurrency.lockutils [req-979910f7-e6a3-4030-9579-441bb6649a01 req-9867ff8f-a0b3-47a3-9cc7-8e1e12613b2d service nova] Releasing lock "refresh_cache-439ee94c-04f2-45d3-a486-81a216c8db4e" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1058.800753] env[63202]: DEBUG oslo_vmware.api [None req-707e40ae-6044-442d-9a89-69e1127b3e4d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385856, 'name': ReconfigVM_Task, 'duration_secs': 0.260889} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.800753] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-707e40ae-6044-442d-9a89-69e1127b3e4d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Reconfigured VM instance instance-00000051 to detach disk 2001 {{(pid=63202) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1058.805190] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-793db438-c123-4f96-9b7d-6a29c0ba39af {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.820676] env[63202]: DEBUG oslo_vmware.api [None req-707e40ae-6044-442d-9a89-69e1127b3e4d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 1058.820676] env[63202]: value = "task-1385859" [ 1058.820676] env[63202]: _type = "Task" [ 1058.820676] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.830441] env[63202]: DEBUG oslo_vmware.api [None req-707e40ae-6044-442d-9a89-69e1127b3e4d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385859, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.836142] env[63202]: DEBUG oslo_vmware.api [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52ae6816-51c4-c55a-a47b-15d21b660691, 'name': SearchDatastore_Task, 'duration_secs': 0.014757} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.836881] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f756ae81-75a5-44c4-8414-3c9223160c4d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.841694] env[63202]: DEBUG oslo_vmware.api [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 1058.841694] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52537a27-7337-4961-0e17-a776c6755e1c" [ 1058.841694] env[63202]: _type = "Task" [ 1058.841694] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.849304] env[63202]: DEBUG oslo_vmware.api [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52537a27-7337-4961-0e17-a776c6755e1c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.897755] env[63202]: DEBUG oslo_vmware.api [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385858, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.118966] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ee02054-3fb6-497e-ad5e-fb06a32df79c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.129635] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-161d635a-13ea-4957-9d76-5ff428682fc4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.161459] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c63368e4-aff4-48ea-a395-3a82a109e34e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.170105] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6c83ce8-eb9b-470f-9851-07c170dc7759 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.185130] env[63202]: DEBUG nova.compute.provider_tree [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1059.335762] env[63202]: DEBUG oslo_vmware.api [None req-707e40ae-6044-442d-9a89-69e1127b3e4d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385859, 'name': ReconfigVM_Task, 'duration_secs': 0.128668} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.338291] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-707e40ae-6044-442d-9a89-69e1127b3e4d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294189', 'volume_id': 'd296fff5-0fd4-4820-8a58-c19b5b2d0da9', 'name': 'volume-d296fff5-0fd4-4820-8a58-c19b5b2d0da9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7437595c-fa35-483e-95f3-b75405b6bd13', 'attached_at': '', 'detached_at': '', 'volume_id': 'd296fff5-0fd4-4820-8a58-c19b5b2d0da9', 'serial': 'd296fff5-0fd4-4820-8a58-c19b5b2d0da9'} {{(pid=63202) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1059.354372] env[63202]: DEBUG oslo_vmware.api [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52537a27-7337-4961-0e17-a776c6755e1c, 'name': SearchDatastore_Task, 'duration_secs': 0.015088} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.354573] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.355078] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] 439ee94c-04f2-45d3-a486-81a216c8db4e/439ee94c-04f2-45d3-a486-81a216c8db4e.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1059.355562] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-02109b8e-abb4-4cf2-867e-37061e3ae874 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.364948] env[63202]: DEBUG oslo_vmware.api [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 1059.364948] env[63202]: value = "task-1385860" [ 1059.364948] env[63202]: _type = "Task" [ 1059.364948] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.376445] env[63202]: DEBUG oslo_vmware.api [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385860, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.400573] env[63202]: DEBUG oslo_vmware.api [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385858, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.688942] env[63202]: DEBUG nova.scheduler.client.report [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1059.880790] env[63202]: DEBUG oslo_vmware.api [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385860, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.902207] env[63202]: DEBUG oslo_vmware.api [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385858, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.912860] env[63202]: DEBUG nova.objects.instance [None req-707e40ae-6044-442d-9a89-69e1127b3e4d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lazy-loading 'flavor' on Instance uuid 7437595c-fa35-483e-95f3-b75405b6bd13 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1060.025046] env[63202]: DEBUG nova.compute.manager [req-e4f1d2eb-b804-40d4-ad2d-99bc352e2bcc req-1a9ee4f6-f474-4a0e-b611-19e0644ae5fb service nova] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Received event network-changed-b78164a7-4413-4943-b4fd-0a55d84f87fe {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1060.025496] env[63202]: DEBUG nova.compute.manager [req-e4f1d2eb-b804-40d4-ad2d-99bc352e2bcc req-1a9ee4f6-f474-4a0e-b611-19e0644ae5fb service nova] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Refreshing instance network info cache due to event network-changed-b78164a7-4413-4943-b4fd-0a55d84f87fe. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1060.025939] env[63202]: DEBUG oslo_concurrency.lockutils [req-e4f1d2eb-b804-40d4-ad2d-99bc352e2bcc req-1a9ee4f6-f474-4a0e-b611-19e0644ae5fb service nova] Acquiring lock "refresh_cache-1123a12e-5218-415d-b286-2f005fe57b29" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.026364] env[63202]: DEBUG oslo_concurrency.lockutils [req-e4f1d2eb-b804-40d4-ad2d-99bc352e2bcc req-1a9ee4f6-f474-4a0e-b611-19e0644ae5fb service nova] Acquired lock "refresh_cache-1123a12e-5218-415d-b286-2f005fe57b29" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.026681] env[63202]: DEBUG nova.network.neutron [req-e4f1d2eb-b804-40d4-ad2d-99bc352e2bcc req-1a9ee4f6-f474-4a0e-b611-19e0644ae5fb service nova] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Refreshing network info cache for port b78164a7-4413-4943-b4fd-0a55d84f87fe {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1060.198288] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.191s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.380185] env[63202]: DEBUG oslo_vmware.api [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385860, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.405191] env[63202]: DEBUG oslo_vmware.api [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385858, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.701600] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Acquiring lock "43659b88-f9d3-49cf-bef3-f8dd14144dc0" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1060.701898] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Lock "43659b88-f9d3-49cf-bef3-f8dd14144dc0" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1060.878986] env[63202]: DEBUG oslo_vmware.api [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385860, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.890676] env[63202]: DEBUG nova.network.neutron [req-e4f1d2eb-b804-40d4-ad2d-99bc352e2bcc req-1a9ee4f6-f474-4a0e-b611-19e0644ae5fb service nova] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Updated VIF entry in instance network info cache for port b78164a7-4413-4943-b4fd-0a55d84f87fe. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1060.891065] env[63202]: DEBUG nova.network.neutron [req-e4f1d2eb-b804-40d4-ad2d-99bc352e2bcc req-1a9ee4f6-f474-4a0e-b611-19e0644ae5fb service nova] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Updating instance_info_cache with network_info: [{"id": "b78164a7-4413-4943-b4fd-0a55d84f87fe", "address": "fa:16:3e:e4:94:7c", "network": {"id": "d36e22d7-c866-4005-8049-ebce7c0376bd", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1130666168-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "9c1a31c777da417c96b2552ca50dc3a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e614f8e-6b11-4b6b-a421-904bca6acd91", "external-id": "nsx-vlan-transportzone-923", "segmentation_id": 923, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb78164a7-44", "ovs_interfaceid": "b78164a7-4413-4943-b4fd-0a55d84f87fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.902711] env[63202]: DEBUG oslo_vmware.api [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385858, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.921847] env[63202]: DEBUG oslo_concurrency.lockutils [None req-707e40ae-6044-442d-9a89-69e1127b3e4d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "7437595c-fa35-483e-95f3-b75405b6bd13" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.294s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.204509] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Lock "43659b88-f9d3-49cf-bef3-f8dd14144dc0" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.502s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.204912] env[63202]: DEBUG nova.compute.manager [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1061.253248] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1061.253584] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1061.380339] env[63202]: DEBUG oslo_vmware.api [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385860, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.397671] env[63202]: DEBUG oslo_concurrency.lockutils [req-e4f1d2eb-b804-40d4-ad2d-99bc352e2bcc req-1a9ee4f6-f474-4a0e-b611-19e0644ae5fb service nova] Releasing lock "refresh_cache-1123a12e-5218-415d-b286-2f005fe57b29" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1061.404256] env[63202]: DEBUG oslo_vmware.api [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385858, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.711027] env[63202]: DEBUG nova.compute.utils [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1061.711410] env[63202]: DEBUG nova.compute.manager [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1061.711893] env[63202]: DEBUG nova.network.neutron [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1061.810168] env[63202]: DEBUG nova.policy [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '02f6449267784ba1a50ca6a9148821b8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b0fe5199b8b940e0b23c49026fdcc685', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 1061.881028] env[63202]: DEBUG oslo_vmware.api [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385860, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.185467} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.881028] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] 439ee94c-04f2-45d3-a486-81a216c8db4e/439ee94c-04f2-45d3-a486-81a216c8db4e.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1061.881028] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1061.881028] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4cebb5a8-9c18-48d9-a6b7-5c17b3d60ee8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.887079] env[63202]: DEBUG oslo_vmware.api [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 1061.887079] env[63202]: value = "task-1385861" [ 1061.887079] env[63202]: _type = "Task" [ 1061.887079] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.895503] env[63202]: DEBUG oslo_vmware.api [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385861, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.903976] env[63202]: DEBUG oslo_vmware.api [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385858, 'name': MoveVirtualDisk_Task, 'duration_secs': 3.148552} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.906903] env[63202]: INFO nova.virt.vmwareapi.ds_util [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_0b4c512d-eb92-4472-b3c8-e79a01cd4464/OSTACK_IMG_0b4c512d-eb92-4472-b3c8-e79a01cd4464.vmdk to [datastore1] devstack-image-cache_base/3db9d5e2-204c-47e7-ae45-3ed10283296d/3db9d5e2-204c-47e7-ae45-3ed10283296d.vmdk. [ 1061.906903] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Cleaning up location [datastore1] OSTACK_IMG_0b4c512d-eb92-4472-b3c8-e79a01cd4464 {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1061.906903] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_0b4c512d-eb92-4472-b3c8-e79a01cd4464 {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1061.906903] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-96a0aaba-1f53-4159-92c8-0c55000be5b8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.913016] env[63202]: DEBUG oslo_vmware.api [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1061.913016] env[63202]: value = "task-1385862" [ 1061.913016] env[63202]: _type = "Task" [ 1061.913016] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.919337] env[63202]: DEBUG oslo_vmware.api [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385862, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.929138] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2ef4997b-172c-4d17-b6cd-0f645eb38a09 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquiring lock "7437595c-fa35-483e-95f3-b75405b6bd13" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.929138] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2ef4997b-172c-4d17-b6cd-0f645eb38a09 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "7437595c-fa35-483e-95f3-b75405b6bd13" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.929138] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2ef4997b-172c-4d17-b6cd-0f645eb38a09 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquiring lock "7437595c-fa35-483e-95f3-b75405b6bd13-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.929138] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2ef4997b-172c-4d17-b6cd-0f645eb38a09 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "7437595c-fa35-483e-95f3-b75405b6bd13-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.929554] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2ef4997b-172c-4d17-b6cd-0f645eb38a09 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "7437595c-fa35-483e-95f3-b75405b6bd13-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.930094] env[63202]: INFO nova.compute.manager [None req-2ef4997b-172c-4d17-b6cd-0f645eb38a09 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Terminating instance [ 1061.932014] env[63202]: DEBUG nova.compute.manager [None req-2ef4997b-172c-4d17-b6cd-0f645eb38a09 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1061.932332] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-2ef4997b-172c-4d17-b6cd-0f645eb38a09 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1061.933232] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-197cf9be-bf2e-4dff-a13f-c19a43bc0bf4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.940660] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ef4997b-172c-4d17-b6cd-0f645eb38a09 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1061.941047] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-69250073-f81a-408a-a973-5a7d0975ddd4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.947944] env[63202]: DEBUG oslo_vmware.api [None req-2ef4997b-172c-4d17-b6cd-0f645eb38a09 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 1061.947944] env[63202]: value = "task-1385863" [ 1061.947944] env[63202]: _type = "Task" [ 1061.947944] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.957248] env[63202]: DEBUG oslo_vmware.api [None req-2ef4997b-172c-4d17-b6cd-0f645eb38a09 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385863, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.193326] env[63202]: DEBUG nova.compute.manager [req-6332eb89-5ef6-4eb1-8cf9-79aca39832a8 req-5f999725-b449-4ef7-9efc-ccdf3138a8d3 service nova] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Received event network-changed-b78164a7-4413-4943-b4fd-0a55d84f87fe {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1062.193326] env[63202]: DEBUG nova.compute.manager [req-6332eb89-5ef6-4eb1-8cf9-79aca39832a8 req-5f999725-b449-4ef7-9efc-ccdf3138a8d3 service nova] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Refreshing instance network info cache due to event network-changed-b78164a7-4413-4943-b4fd-0a55d84f87fe. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1062.193326] env[63202]: DEBUG oslo_concurrency.lockutils [req-6332eb89-5ef6-4eb1-8cf9-79aca39832a8 req-5f999725-b449-4ef7-9efc-ccdf3138a8d3 service nova] Acquiring lock "refresh_cache-1123a12e-5218-415d-b286-2f005fe57b29" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1062.193326] env[63202]: DEBUG oslo_concurrency.lockutils [req-6332eb89-5ef6-4eb1-8cf9-79aca39832a8 req-5f999725-b449-4ef7-9efc-ccdf3138a8d3 service nova] Acquired lock "refresh_cache-1123a12e-5218-415d-b286-2f005fe57b29" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1062.193326] env[63202]: DEBUG nova.network.neutron [req-6332eb89-5ef6-4eb1-8cf9-79aca39832a8 req-5f999725-b449-4ef7-9efc-ccdf3138a8d3 service nova] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Refreshing network info cache for port b78164a7-4413-4943-b4fd-0a55d84f87fe {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1062.216042] env[63202]: DEBUG nova.compute.manager [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1062.359723] env[63202]: DEBUG nova.network.neutron [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Successfully created port: aaf1e261-e958-4575-a0c8-b9959bf1bd4d {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1062.398424] env[63202]: DEBUG oslo_vmware.api [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385861, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067805} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.398717] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1062.399787] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a1d0e5d-0d94-4933-b168-2702f5e6947d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.424496] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] 439ee94c-04f2-45d3-a486-81a216c8db4e/439ee94c-04f2-45d3-a486-81a216c8db4e.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1062.428527] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7304e588-8b8e-4e7d-bc90-fb097a2e0e53 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.450803] env[63202]: DEBUG oslo_vmware.api [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385862, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141546} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.455228] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1062.455782] env[63202]: DEBUG oslo_concurrency.lockutils [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3db9d5e2-204c-47e7-ae45-3ed10283296d/3db9d5e2-204c-47e7-ae45-3ed10283296d.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1062.455782] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3db9d5e2-204c-47e7-ae45-3ed10283296d/3db9d5e2-204c-47e7-ae45-3ed10283296d.vmdk to [datastore1] 8437f856-a707-49c5-b8eb-5a22cdb990f8/8437f856-a707-49c5-b8eb-5a22cdb990f8.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1062.456048] env[63202]: DEBUG oslo_vmware.api [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 1062.456048] env[63202]: value = "task-1385864" [ 1062.456048] env[63202]: _type = "Task" [ 1062.456048] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.456640] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1d6b5848-ef91-4e9b-b9c5-203ef1d30792 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.465634] env[63202]: DEBUG oslo_vmware.api [None req-2ef4997b-172c-4d17-b6cd-0f645eb38a09 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385863, 'name': PowerOffVM_Task, 'duration_secs': 0.199477} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.466157] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ef4997b-172c-4d17-b6cd-0f645eb38a09 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1062.466339] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-2ef4997b-172c-4d17-b6cd-0f645eb38a09 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1062.466594] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-06b05c9b-0184-4969-8f5c-c628f7c08923 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.472883] env[63202]: DEBUG oslo_vmware.api [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385864, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.474214] env[63202]: DEBUG oslo_vmware.api [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1062.474214] env[63202]: value = "task-1385865" [ 1062.474214] env[63202]: _type = "Task" [ 1062.474214] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.482791] env[63202]: DEBUG oslo_vmware.api [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385865, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.535318] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-2ef4997b-172c-4d17-b6cd-0f645eb38a09 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1062.535568] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-2ef4997b-172c-4d17-b6cd-0f645eb38a09 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Deleting contents of the VM from datastore datastore2 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1062.535804] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ef4997b-172c-4d17-b6cd-0f645eb38a09 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Deleting the datastore file [datastore2] 7437595c-fa35-483e-95f3-b75405b6bd13 {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1062.536101] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-83d48468-c367-49d0-afb2-801e8e7d2988 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.543887] env[63202]: DEBUG oslo_vmware.api [None req-2ef4997b-172c-4d17-b6cd-0f645eb38a09 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 1062.543887] env[63202]: value = "task-1385867" [ 1062.543887] env[63202]: _type = "Task" [ 1062.543887] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.552354] env[63202]: DEBUG oslo_vmware.api [None req-2ef4997b-172c-4d17-b6cd-0f645eb38a09 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385867, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.972232] env[63202]: DEBUG oslo_vmware.api [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385864, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.983504] env[63202]: DEBUG oslo_vmware.api [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385865, 'name': CopyVirtualDisk_Task} progress is 12%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.049051] env[63202]: DEBUG nova.network.neutron [req-6332eb89-5ef6-4eb1-8cf9-79aca39832a8 req-5f999725-b449-4ef7-9efc-ccdf3138a8d3 service nova] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Updated VIF entry in instance network info cache for port b78164a7-4413-4943-b4fd-0a55d84f87fe. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1063.049514] env[63202]: DEBUG nova.network.neutron [req-6332eb89-5ef6-4eb1-8cf9-79aca39832a8 req-5f999725-b449-4ef7-9efc-ccdf3138a8d3 service nova] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Updating instance_info_cache with network_info: [{"id": "b78164a7-4413-4943-b4fd-0a55d84f87fe", "address": "fa:16:3e:e4:94:7c", "network": {"id": "d36e22d7-c866-4005-8049-ebce7c0376bd", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1130666168-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.169", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "9c1a31c777da417c96b2552ca50dc3a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e614f8e-6b11-4b6b-a421-904bca6acd91", "external-id": "nsx-vlan-transportzone-923", "segmentation_id": 923, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb78164a7-44", "ovs_interfaceid": "b78164a7-4413-4943-b4fd-0a55d84f87fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1063.055922] env[63202]: DEBUG oslo_vmware.api [None req-2ef4997b-172c-4d17-b6cd-0f645eb38a09 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385867, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.335985} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.056190] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ef4997b-172c-4d17-b6cd-0f645eb38a09 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1063.056462] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-2ef4997b-172c-4d17-b6cd-0f645eb38a09 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Deleted contents of the VM from datastore datastore2 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1063.056611] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-2ef4997b-172c-4d17-b6cd-0f645eb38a09 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1063.056726] env[63202]: INFO nova.compute.manager [None req-2ef4997b-172c-4d17-b6cd-0f645eb38a09 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1063.056969] env[63202]: DEBUG oslo.service.loopingcall [None req-2ef4997b-172c-4d17-b6cd-0f645eb38a09 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1063.057198] env[63202]: DEBUG nova.compute.manager [-] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1063.057277] env[63202]: DEBUG nova.network.neutron [-] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1063.227570] env[63202]: DEBUG nova.compute.manager [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1063.253871] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager.update_available_resource {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1063.263177] env[63202]: DEBUG nova.virt.hardware [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1063.263457] env[63202]: DEBUG nova.virt.hardware [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1063.263593] env[63202]: DEBUG nova.virt.hardware [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1063.263785] env[63202]: DEBUG nova.virt.hardware [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1063.264052] env[63202]: DEBUG nova.virt.hardware [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1063.264171] env[63202]: DEBUG nova.virt.hardware [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1063.264333] env[63202]: DEBUG nova.virt.hardware [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1063.264477] env[63202]: DEBUG nova.virt.hardware [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1063.264755] env[63202]: DEBUG nova.virt.hardware [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1063.264914] env[63202]: DEBUG nova.virt.hardware [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1063.265066] env[63202]: DEBUG nova.virt.hardware [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1063.266116] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd6bdaf5-0921-4209-a9bc-5b20982b29d9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.276515] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfccd69e-ffc9-441f-b698-9aac1575bffd {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.475777] env[63202]: DEBUG oslo_vmware.api [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385864, 'name': ReconfigVM_Task, 'duration_secs': 0.609464} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.480253] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Reconfigured VM instance instance-00000064 to attach disk [datastore1] 439ee94c-04f2-45d3-a486-81a216c8db4e/439ee94c-04f2-45d3-a486-81a216c8db4e.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1063.480978] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-540f3500-3ec1-4b15-b223-5f97c81a8092 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.491540] env[63202]: DEBUG oslo_vmware.api [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385865, 'name': CopyVirtualDisk_Task} progress is 32%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.495492] env[63202]: DEBUG oslo_vmware.api [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 1063.495492] env[63202]: value = "task-1385868" [ 1063.495492] env[63202]: _type = "Task" [ 1063.495492] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.512154] env[63202]: DEBUG oslo_vmware.api [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385868, 'name': Rename_Task} progress is 10%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.552712] env[63202]: DEBUG oslo_concurrency.lockutils [req-6332eb89-5ef6-4eb1-8cf9-79aca39832a8 req-5f999725-b449-4ef7-9efc-ccdf3138a8d3 service nova] Releasing lock "refresh_cache-1123a12e-5218-415d-b286-2f005fe57b29" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.758803] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.759094] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.759243] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.759406] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63202) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1063.760470] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1f44449-df5b-4fd2-a5e1-df2a44a39547 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.773571] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-002c6d1d-0045-4d48-bb24-ff30a502561b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.791066] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-568e738a-0d7a-49e4-b96e-48682d6639ae {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.801628] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7ce371a-561c-47b4-9def-5718584ea4f4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.838545] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180238MB free_disk=176GB free_vcpus=48 pci_devices=None {{(pid=63202) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1063.838703] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.838912] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.861526] env[63202]: DEBUG nova.compute.manager [req-3dbfdcf7-fb7b-4862-9253-cbbc33cb44a0 req-8df974b8-4dad-4910-be50-8fe21407bfa2 service nova] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Received event network-vif-deleted-5eeb4ae8-9356-402a-965d-224788f04129 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1063.861669] env[63202]: INFO nova.compute.manager [req-3dbfdcf7-fb7b-4862-9253-cbbc33cb44a0 req-8df974b8-4dad-4910-be50-8fe21407bfa2 service nova] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Neutron deleted interface 5eeb4ae8-9356-402a-965d-224788f04129; detaching it from the instance and deleting it from the info cache [ 1063.862949] env[63202]: DEBUG nova.network.neutron [req-3dbfdcf7-fb7b-4862-9253-cbbc33cb44a0 req-8df974b8-4dad-4910-be50-8fe21407bfa2 service nova] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1063.990533] env[63202]: DEBUG oslo_vmware.api [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385865, 'name': CopyVirtualDisk_Task} progress is 52%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.009392] env[63202]: DEBUG oslo_vmware.api [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385868, 'name': Rename_Task, 'duration_secs': 0.394422} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.009721] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1064.009992] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-493f6012-c6cc-4aba-93c9-55faa29e030c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.018408] env[63202]: DEBUG oslo_vmware.api [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 1064.018408] env[63202]: value = "task-1385869" [ 1064.018408] env[63202]: _type = "Task" [ 1064.018408] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.031855] env[63202]: DEBUG oslo_vmware.api [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385869, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.347474] env[63202]: DEBUG nova.network.neutron [-] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1064.366756] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0fe123c6-efd4-4698-aad9-0472fdf1b75b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.384124] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf52cc45-b750-4022-a4b8-89a05c9d8abb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.414144] env[63202]: DEBUG nova.compute.manager [req-3dbfdcf7-fb7b-4862-9253-cbbc33cb44a0 req-8df974b8-4dad-4910-be50-8fe21407bfa2 service nova] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Detach interface failed, port_id=5eeb4ae8-9356-402a-965d-224788f04129, reason: Instance 7437595c-fa35-483e-95f3-b75405b6bd13 could not be found. {{(pid=63202) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1064.487799] env[63202]: DEBUG oslo_vmware.api [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385865, 'name': CopyVirtualDisk_Task} progress is 74%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.530325] env[63202]: DEBUG oslo_vmware.api [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385869, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.772686] env[63202]: DEBUG nova.network.neutron [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Successfully updated port: aaf1e261-e958-4575-a0c8-b9959bf1bd4d {{(pid=63202) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1064.782647] env[63202]: DEBUG nova.compute.manager [req-cffd8f7d-0cc4-4de1-b7d7-e1eb9075c5f6 req-e74fa45c-6710-4775-9019-b098955ca441 service nova] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Received event network-changed-b78164a7-4413-4943-b4fd-0a55d84f87fe {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1064.782907] env[63202]: DEBUG nova.compute.manager [req-cffd8f7d-0cc4-4de1-b7d7-e1eb9075c5f6 req-e74fa45c-6710-4775-9019-b098955ca441 service nova] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Refreshing instance network info cache due to event network-changed-b78164a7-4413-4943-b4fd-0a55d84f87fe. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1064.783141] env[63202]: DEBUG oslo_concurrency.lockutils [req-cffd8f7d-0cc4-4de1-b7d7-e1eb9075c5f6 req-e74fa45c-6710-4775-9019-b098955ca441 service nova] Acquiring lock "refresh_cache-1123a12e-5218-415d-b286-2f005fe57b29" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1064.783285] env[63202]: DEBUG oslo_concurrency.lockutils [req-cffd8f7d-0cc4-4de1-b7d7-e1eb9075c5f6 req-e74fa45c-6710-4775-9019-b098955ca441 service nova] Acquired lock "refresh_cache-1123a12e-5218-415d-b286-2f005fe57b29" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.783462] env[63202]: DEBUG nova.network.neutron [req-cffd8f7d-0cc4-4de1-b7d7-e1eb9075c5f6 req-e74fa45c-6710-4775-9019-b098955ca441 service nova] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Refreshing network info cache for port b78164a7-4413-4943-b4fd-0a55d84f87fe {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1064.854197] env[63202]: INFO nova.compute.manager [-] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Took 1.80 seconds to deallocate network for instance. [ 1064.880779] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 7437595c-fa35-483e-95f3-b75405b6bd13 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1064.880938] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance e1e2e8a4-80aa-45eb-a90e-d20cba2943ee actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1064.881095] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 8437f856-a707-49c5-b8eb-5a22cdb990f8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1064.881216] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 1123a12e-5218-415d-b286-2f005fe57b29 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1064.881331] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 439ee94c-04f2-45d3-a486-81a216c8db4e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1064.881482] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 6e172a2a-1918-430c-84ce-005fef97bbf1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1064.881673] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=63202) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1064.881813] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=63202) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1064.986819] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eeef6ff-2a49-4bd7-b0ab-62827e2be98d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.995407] env[63202]: DEBUG oslo_vmware.api [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385865, 'name': CopyVirtualDisk_Task} progress is 91%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.000251] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4de6a739-3b5c-4e7e-a180-f01b002ac77b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.036410] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83d54869-e886-4367-81c4-ce3a34bc19e2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.045670] env[63202]: DEBUG oslo_vmware.api [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385869, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.049594] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf1e228b-69e2-47da-86e2-f8d955d2f4db {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.065840] env[63202]: DEBUG nova.compute.provider_tree [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1065.277929] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Acquiring lock "refresh_cache-6e172a2a-1918-430c-84ce-005fef97bbf1" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1065.279585] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Acquired lock "refresh_cache-6e172a2a-1918-430c-84ce-005fef97bbf1" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.279585] env[63202]: DEBUG nova.network.neutron [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1065.367820] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2ef4997b-172c-4d17-b6cd-0f645eb38a09 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.451472] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Acquiring lock "9646810e-06d2-4de0-8f0c-e8719541da53" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.451788] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Lock "9646810e-06d2-4de0-8f0c-e8719541da53" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.487848] env[63202]: DEBUG oslo_vmware.api [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385865, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.79506} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.488772] env[63202]: DEBUG nova.network.neutron [req-cffd8f7d-0cc4-4de1-b7d7-e1eb9075c5f6 req-e74fa45c-6710-4775-9019-b098955ca441 service nova] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Updated VIF entry in instance network info cache for port b78164a7-4413-4943-b4fd-0a55d84f87fe. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1065.489266] env[63202]: DEBUG nova.network.neutron [req-cffd8f7d-0cc4-4de1-b7d7-e1eb9075c5f6 req-e74fa45c-6710-4775-9019-b098955ca441 service nova] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Updating instance_info_cache with network_info: [{"id": "b78164a7-4413-4943-b4fd-0a55d84f87fe", "address": "fa:16:3e:e4:94:7c", "network": {"id": "d36e22d7-c866-4005-8049-ebce7c0376bd", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1130666168-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "9c1a31c777da417c96b2552ca50dc3a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e614f8e-6b11-4b6b-a421-904bca6acd91", "external-id": "nsx-vlan-transportzone-923", "segmentation_id": 923, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb78164a7-44", "ovs_interfaceid": "b78164a7-4413-4943-b4fd-0a55d84f87fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1065.490838] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3db9d5e2-204c-47e7-ae45-3ed10283296d/3db9d5e2-204c-47e7-ae45-3ed10283296d.vmdk to [datastore1] 8437f856-a707-49c5-b8eb-5a22cdb990f8/8437f856-a707-49c5-b8eb-5a22cdb990f8.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1065.492362] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc201102-c8bf-401c-be1e-ca84b96e9d05 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.517019] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] 8437f856-a707-49c5-b8eb-5a22cdb990f8/8437f856-a707-49c5-b8eb-5a22cdb990f8.vmdk or device None with type streamOptimized {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1065.517823] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-17fd0bbe-a4b5-4b73-b223-77e6ba03303f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.540849] env[63202]: DEBUG oslo_vmware.api [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1065.540849] env[63202]: value = "task-1385870" [ 1065.540849] env[63202]: _type = "Task" [ 1065.540849] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.545335] env[63202]: DEBUG oslo_vmware.api [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385869, 'name': PowerOnVM_Task, 'duration_secs': 1.134642} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.549668] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1065.549887] env[63202]: INFO nova.compute.manager [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Took 10.62 seconds to spawn the instance on the hypervisor. [ 1065.550082] env[63202]: DEBUG nova.compute.manager [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1065.551085] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-960344ca-a465-4847-99c4-7a7f19170534 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.558580] env[63202]: DEBUG oslo_vmware.api [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385870, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.568428] env[63202]: DEBUG nova.scheduler.client.report [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1065.681940] env[63202]: DEBUG oslo_concurrency.lockutils [None req-fe0ebb9d-359a-44f5-9fe9-5ec465a84793 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Acquiring lock "1123a12e-5218-415d-b286-2f005fe57b29" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.682251] env[63202]: DEBUG oslo_concurrency.lockutils [None req-fe0ebb9d-359a-44f5-9fe9-5ec465a84793 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Lock "1123a12e-5218-415d-b286-2f005fe57b29" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.682471] env[63202]: DEBUG oslo_concurrency.lockutils [None req-fe0ebb9d-359a-44f5-9fe9-5ec465a84793 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Acquiring lock "1123a12e-5218-415d-b286-2f005fe57b29-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.682716] env[63202]: DEBUG oslo_concurrency.lockutils [None req-fe0ebb9d-359a-44f5-9fe9-5ec465a84793 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Lock "1123a12e-5218-415d-b286-2f005fe57b29-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.682876] env[63202]: DEBUG oslo_concurrency.lockutils [None req-fe0ebb9d-359a-44f5-9fe9-5ec465a84793 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Lock "1123a12e-5218-415d-b286-2f005fe57b29-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.686870] env[63202]: INFO nova.compute.manager [None req-fe0ebb9d-359a-44f5-9fe9-5ec465a84793 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Terminating instance [ 1065.688824] env[63202]: DEBUG nova.compute.manager [None req-fe0ebb9d-359a-44f5-9fe9-5ec465a84793 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1065.689066] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-fe0ebb9d-359a-44f5-9fe9-5ec465a84793 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1065.689990] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02637ef6-90bd-4158-8ddd-b0923e13cefa {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.698801] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe0ebb9d-359a-44f5-9fe9-5ec465a84793 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1065.699089] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-77ded560-320c-4594-aa9f-5ef40b76441c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.706881] env[63202]: DEBUG oslo_vmware.api [None req-fe0ebb9d-359a-44f5-9fe9-5ec465a84793 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Waiting for the task: (returnval){ [ 1065.706881] env[63202]: value = "task-1385871" [ 1065.706881] env[63202]: _type = "Task" [ 1065.706881] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.716090] env[63202]: DEBUG oslo_vmware.api [None req-fe0ebb9d-359a-44f5-9fe9-5ec465a84793 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Task: {'id': task-1385871, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.814650] env[63202]: DEBUG nova.network.neutron [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1065.954743] env[63202]: DEBUG nova.compute.manager [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1065.982374] env[63202]: DEBUG nova.network.neutron [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Updating instance_info_cache with network_info: [{"id": "aaf1e261-e958-4575-a0c8-b9959bf1bd4d", "address": "fa:16:3e:96:4d:5a", "network": {"id": "b906037c-ca02-45b3-83f0-620a7a89656c", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-877611281-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b0fe5199b8b940e0b23c49026fdcc685", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48512b02-ad5c-4105-ba7d-fd4775acf8e1", "external-id": "nsx-vlan-transportzone-516", "segmentation_id": 516, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaaf1e261-e9", "ovs_interfaceid": "aaf1e261-e958-4575-a0c8-b9959bf1bd4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1065.993222] env[63202]: DEBUG oslo_concurrency.lockutils [req-cffd8f7d-0cc4-4de1-b7d7-e1eb9075c5f6 req-e74fa45c-6710-4775-9019-b098955ca441 service nova] Releasing lock "refresh_cache-1123a12e-5218-415d-b286-2f005fe57b29" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1065.993540] env[63202]: DEBUG nova.compute.manager [req-cffd8f7d-0cc4-4de1-b7d7-e1eb9075c5f6 req-e74fa45c-6710-4775-9019-b098955ca441 service nova] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Received event network-changed-b78164a7-4413-4943-b4fd-0a55d84f87fe {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1065.993714] env[63202]: DEBUG nova.compute.manager [req-cffd8f7d-0cc4-4de1-b7d7-e1eb9075c5f6 req-e74fa45c-6710-4775-9019-b098955ca441 service nova] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Refreshing instance network info cache due to event network-changed-b78164a7-4413-4943-b4fd-0a55d84f87fe. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1065.993964] env[63202]: DEBUG oslo_concurrency.lockutils [req-cffd8f7d-0cc4-4de1-b7d7-e1eb9075c5f6 req-e74fa45c-6710-4775-9019-b098955ca441 service nova] Acquiring lock "refresh_cache-1123a12e-5218-415d-b286-2f005fe57b29" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1065.994167] env[63202]: DEBUG oslo_concurrency.lockutils [req-cffd8f7d-0cc4-4de1-b7d7-e1eb9075c5f6 req-e74fa45c-6710-4775-9019-b098955ca441 service nova] Acquired lock "refresh_cache-1123a12e-5218-415d-b286-2f005fe57b29" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.994463] env[63202]: DEBUG nova.network.neutron [req-cffd8f7d-0cc4-4de1-b7d7-e1eb9075c5f6 req-e74fa45c-6710-4775-9019-b098955ca441 service nova] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Refreshing network info cache for port b78164a7-4413-4943-b4fd-0a55d84f87fe {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1066.055019] env[63202]: DEBUG oslo_vmware.api [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385870, 'name': ReconfigVM_Task, 'duration_secs': 0.321811} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.055332] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Reconfigured VM instance instance-00000055 to attach disk [datastore1] 8437f856-a707-49c5-b8eb-5a22cdb990f8/8437f856-a707-49c5-b8eb-5a22cdb990f8.vmdk or device None with type streamOptimized {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1066.056825] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'encrypted': False, 'encryption_format': None, 'size': 0, 'device_type': 'disk', 'guest_format': None, 'encryption_options': None, 'encryption_secret_uuid': None, 'device_name': '/dev/sda', 'disk_bus': None, 'boot_index': 0, 'image_id': 'bb172aa4-a5a8-4395-9793-2416c30721cd'}], 'ephemerals': [], 'block_device_mapping': [{'device_type': None, 'guest_format': None, 'mount_device': '/dev/sdb', 'delete_on_termination': False, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294210', 'volume_id': '81b406ab-e721-4271-b298-75cded690f20', 'name': 'volume-81b406ab-e721-4271-b298-75cded690f20', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '8437f856-a707-49c5-b8eb-5a22cdb990f8', 'attached_at': '', 'detached_at': '', 'volume_id': '81b406ab-e721-4271-b298-75cded690f20', 'serial': '81b406ab-e721-4271-b298-75cded690f20'}, 'disk_bus': None, 'boot_index': None, 'attachment_id': 'fdfb419e-c6e2-4a9a-92fd-a47ad5e49c30', 'volume_type': None}], 'swap': None} {{(pid=63202) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1066.057071] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Volume attach. Driver type: vmdk {{(pid=63202) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1066.057275] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294210', 'volume_id': '81b406ab-e721-4271-b298-75cded690f20', 'name': 'volume-81b406ab-e721-4271-b298-75cded690f20', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '8437f856-a707-49c5-b8eb-5a22cdb990f8', 'attached_at': '', 'detached_at': '', 'volume_id': '81b406ab-e721-4271-b298-75cded690f20', 'serial': '81b406ab-e721-4271-b298-75cded690f20'} {{(pid=63202) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1066.058291] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e3bb477-0c4b-423b-a5c2-77caf4669c45 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.078923] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63202) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1066.078923] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.239s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.080206] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2ef4997b-172c-4d17-b6cd-0f645eb38a09 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.712s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.080700] env[63202]: DEBUG nova.objects.instance [None req-2ef4997b-172c-4d17-b6cd-0f645eb38a09 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lazy-loading 'resources' on Instance uuid 7437595c-fa35-483e-95f3-b75405b6bd13 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1066.082311] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23c2f7cc-c511-4e6b-b75e-49195b433aac {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.086014] env[63202]: INFO nova.compute.manager [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Took 17.06 seconds to build instance. [ 1066.122700] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] volume-81b406ab-e721-4271-b298-75cded690f20/volume-81b406ab-e721-4271-b298-75cded690f20.vmdk or device None with type thin {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1066.123837] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0c2d8a0c-f864-4110-9ac9-9b9dc482945f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.151140] env[63202]: DEBUG oslo_vmware.api [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1066.151140] env[63202]: value = "task-1385872" [ 1066.151140] env[63202]: _type = "Task" [ 1066.151140] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.159286] env[63202]: DEBUG oslo_vmware.api [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385872, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.218388] env[63202]: DEBUG oslo_vmware.api [None req-fe0ebb9d-359a-44f5-9fe9-5ec465a84793 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Task: {'id': task-1385871, 'name': PowerOffVM_Task, 'duration_secs': 0.397022} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.218388] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe0ebb9d-359a-44f5-9fe9-5ec465a84793 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1066.218388] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-fe0ebb9d-359a-44f5-9fe9-5ec465a84793 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1066.218388] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4f1ae53f-f381-44f4-b7e8-e93131956dfc {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.336220] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-fe0ebb9d-359a-44f5-9fe9-5ec465a84793 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1066.336460] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-fe0ebb9d-359a-44f5-9fe9-5ec465a84793 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Deleting contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1066.336982] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe0ebb9d-359a-44f5-9fe9-5ec465a84793 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Deleting the datastore file [datastore1] 1123a12e-5218-415d-b286-2f005fe57b29 {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1066.337114] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a9bd6b2e-e745-4262-9ee0-3e770605f68c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.343466] env[63202]: DEBUG oslo_vmware.api [None req-fe0ebb9d-359a-44f5-9fe9-5ec465a84793 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Waiting for the task: (returnval){ [ 1066.343466] env[63202]: value = "task-1385874" [ 1066.343466] env[63202]: _type = "Task" [ 1066.343466] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.351812] env[63202]: DEBUG oslo_vmware.api [None req-fe0ebb9d-359a-44f5-9fe9-5ec465a84793 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Task: {'id': task-1385874, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.477267] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.484958] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Releasing lock "refresh_cache-6e172a2a-1918-430c-84ce-005fef97bbf1" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1066.485316] env[63202]: DEBUG nova.compute.manager [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Instance network_info: |[{"id": "aaf1e261-e958-4575-a0c8-b9959bf1bd4d", "address": "fa:16:3e:96:4d:5a", "network": {"id": "b906037c-ca02-45b3-83f0-620a7a89656c", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-877611281-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b0fe5199b8b940e0b23c49026fdcc685", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48512b02-ad5c-4105-ba7d-fd4775acf8e1", "external-id": "nsx-vlan-transportzone-516", "segmentation_id": 516, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaaf1e261-e9", "ovs_interfaceid": "aaf1e261-e958-4575-a0c8-b9959bf1bd4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1066.485704] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:96:4d:5a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '48512b02-ad5c-4105-ba7d-fd4775acf8e1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'aaf1e261-e958-4575-a0c8-b9959bf1bd4d', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1066.493436] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Creating folder: Project (b0fe5199b8b940e0b23c49026fdcc685). Parent ref: group-v294090. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1066.493692] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2284c1bb-8492-4220-a54a-197bb9f74e77 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.505387] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Created folder: Project (b0fe5199b8b940e0b23c49026fdcc685) in parent group-v294090. [ 1066.506034] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Creating folder: Instances. Parent ref: group-v294218. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1066.506034] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-41f6cce6-850b-42d3-8efa-43bb51eb97e5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.515413] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Created folder: Instances in parent group-v294218. [ 1066.515634] env[63202]: DEBUG oslo.service.loopingcall [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1066.515824] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1066.516059] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b20cd702-6f12-4fee-b769-1451644e6a15 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.537958] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1066.537958] env[63202]: value = "task-1385877" [ 1066.537958] env[63202]: _type = "Task" [ 1066.537958] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.545578] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385877, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.588930] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3b34197c-e4a6-4990-8dd8-91c4e9cb8152 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "439ee94c-04f2-45d3-a486-81a216c8db4e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.574s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.663795] env[63202]: DEBUG oslo_vmware.api [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385872, 'name': ReconfigVM_Task, 'duration_secs': 0.286374} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.664193] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Reconfigured VM instance instance-00000055 to attach disk [datastore1] volume-81b406ab-e721-4271-b298-75cded690f20/volume-81b406ab-e721-4271-b298-75cded690f20.vmdk or device None with type thin {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1066.670916] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-37147ffe-119f-45b4-8afc-33f17c0ebbfb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.683626] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19b683bd-a52a-4f67-9a32-aac2fe28d506 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.687631] env[63202]: DEBUG oslo_vmware.api [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1066.687631] env[63202]: value = "task-1385878" [ 1066.687631] env[63202]: _type = "Task" [ 1066.687631] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.689229] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a31960a5-3c43-4c5a-a5a0-a9585adc3229 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Suspending the VM {{(pid=63202) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1066.690704] env[63202]: DEBUG nova.network.neutron [req-cffd8f7d-0cc4-4de1-b7d7-e1eb9075c5f6 req-e74fa45c-6710-4775-9019-b098955ca441 service nova] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Updated VIF entry in instance network info cache for port b78164a7-4413-4943-b4fd-0a55d84f87fe. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1066.691054] env[63202]: DEBUG nova.network.neutron [req-cffd8f7d-0cc4-4de1-b7d7-e1eb9075c5f6 req-e74fa45c-6710-4775-9019-b098955ca441 service nova] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Updating instance_info_cache with network_info: [{"id": "b78164a7-4413-4943-b4fd-0a55d84f87fe", "address": "fa:16:3e:e4:94:7c", "network": {"id": "d36e22d7-c866-4005-8049-ebce7c0376bd", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1130666168-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "9c1a31c777da417c96b2552ca50dc3a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e614f8e-6b11-4b6b-a421-904bca6acd91", "external-id": "nsx-vlan-transportzone-923", "segmentation_id": 923, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb78164a7-44", "ovs_interfaceid": "b78164a7-4413-4943-b4fd-0a55d84f87fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1066.695887] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-66bb22ad-89e1-49d0-bd8f-e2706e8cfdc6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.697277] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac28d1c6-afcf-4825-8ff6-8a4fc4861f0b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.706105] env[63202]: DEBUG oslo_vmware.api [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385878, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.709409] env[63202]: DEBUG oslo_vmware.api [None req-a31960a5-3c43-4c5a-a5a0-a9585adc3229 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 1066.709409] env[63202]: value = "task-1385879" [ 1066.709409] env[63202]: _type = "Task" [ 1066.709409] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.710770] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f88cae05-dba9-4343-8fc1-2c0a9bca9fb2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.723079] env[63202]: DEBUG oslo_vmware.api [None req-a31960a5-3c43-4c5a-a5a0-a9585adc3229 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385879, 'name': SuspendVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.751466] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d451c799-1697-4e63-b8c8-39f9c42d5789 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.759785] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c4d1c4a-8670-4332-a3b2-a93b7dd29883 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.778905] env[63202]: DEBUG nova.compute.provider_tree [None req-2ef4997b-172c-4d17-b6cd-0f645eb38a09 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1066.814263] env[63202]: DEBUG nova.compute.manager [req-c23354e6-f223-4abd-b570-433e361ad6f0 req-6b328960-156d-481d-8179-910bcaabba1e service nova] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Received event network-vif-plugged-aaf1e261-e958-4575-a0c8-b9959bf1bd4d {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1066.814564] env[63202]: DEBUG oslo_concurrency.lockutils [req-c23354e6-f223-4abd-b570-433e361ad6f0 req-6b328960-156d-481d-8179-910bcaabba1e service nova] Acquiring lock "6e172a2a-1918-430c-84ce-005fef97bbf1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.814717] env[63202]: DEBUG oslo_concurrency.lockutils [req-c23354e6-f223-4abd-b570-433e361ad6f0 req-6b328960-156d-481d-8179-910bcaabba1e service nova] Lock "6e172a2a-1918-430c-84ce-005fef97bbf1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.814864] env[63202]: DEBUG oslo_concurrency.lockutils [req-c23354e6-f223-4abd-b570-433e361ad6f0 req-6b328960-156d-481d-8179-910bcaabba1e service nova] Lock "6e172a2a-1918-430c-84ce-005fef97bbf1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.815205] env[63202]: DEBUG nova.compute.manager [req-c23354e6-f223-4abd-b570-433e361ad6f0 req-6b328960-156d-481d-8179-910bcaabba1e service nova] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] No waiting events found dispatching network-vif-plugged-aaf1e261-e958-4575-a0c8-b9959bf1bd4d {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1066.815205] env[63202]: WARNING nova.compute.manager [req-c23354e6-f223-4abd-b570-433e361ad6f0 req-6b328960-156d-481d-8179-910bcaabba1e service nova] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Received unexpected event network-vif-plugged-aaf1e261-e958-4575-a0c8-b9959bf1bd4d for instance with vm_state building and task_state spawning. [ 1066.815467] env[63202]: DEBUG nova.compute.manager [req-c23354e6-f223-4abd-b570-433e361ad6f0 req-6b328960-156d-481d-8179-910bcaabba1e service nova] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Received event network-changed-aaf1e261-e958-4575-a0c8-b9959bf1bd4d {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1066.815467] env[63202]: DEBUG nova.compute.manager [req-c23354e6-f223-4abd-b570-433e361ad6f0 req-6b328960-156d-481d-8179-910bcaabba1e service nova] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Refreshing instance network info cache due to event network-changed-aaf1e261-e958-4575-a0c8-b9959bf1bd4d. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1066.815653] env[63202]: DEBUG oslo_concurrency.lockutils [req-c23354e6-f223-4abd-b570-433e361ad6f0 req-6b328960-156d-481d-8179-910bcaabba1e service nova] Acquiring lock "refresh_cache-6e172a2a-1918-430c-84ce-005fef97bbf1" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1066.815850] env[63202]: DEBUG oslo_concurrency.lockutils [req-c23354e6-f223-4abd-b570-433e361ad6f0 req-6b328960-156d-481d-8179-910bcaabba1e service nova] Acquired lock "refresh_cache-6e172a2a-1918-430c-84ce-005fef97bbf1" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1066.815935] env[63202]: DEBUG nova.network.neutron [req-c23354e6-f223-4abd-b570-433e361ad6f0 req-6b328960-156d-481d-8179-910bcaabba1e service nova] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Refreshing network info cache for port aaf1e261-e958-4575-a0c8-b9959bf1bd4d {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1066.853990] env[63202]: DEBUG oslo_vmware.api [None req-fe0ebb9d-359a-44f5-9fe9-5ec465a84793 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Task: {'id': task-1385874, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.215761} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.854276] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe0ebb9d-359a-44f5-9fe9-5ec465a84793 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1066.854469] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-fe0ebb9d-359a-44f5-9fe9-5ec465a84793 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Deleted contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1066.854672] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-fe0ebb9d-359a-44f5-9fe9-5ec465a84793 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1066.854816] env[63202]: INFO nova.compute.manager [None req-fe0ebb9d-359a-44f5-9fe9-5ec465a84793 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1066.855094] env[63202]: DEBUG oslo.service.loopingcall [None req-fe0ebb9d-359a-44f5-9fe9-5ec465a84793 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1066.855305] env[63202]: DEBUG nova.compute.manager [-] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1066.855400] env[63202]: DEBUG nova.network.neutron [-] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1067.048304] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385877, 'name': CreateVM_Task, 'duration_secs': 0.359311} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.048505] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1067.049181] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1067.049489] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.049688] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1067.050025] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4a3bc1a-8fa3-4e04-8296-1366619ce9a0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.054911] env[63202]: DEBUG oslo_vmware.api [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Waiting for the task: (returnval){ [ 1067.054911] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52877c3a-a745-d108-abc7-b75a187ba919" [ 1067.054911] env[63202]: _type = "Task" [ 1067.054911] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.062734] env[63202]: DEBUG oslo_vmware.api [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52877c3a-a745-d108-abc7-b75a187ba919, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.075763] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1067.076037] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1067.076169] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1067.076331] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1067.076493] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1067.076630] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63202) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1067.200090] env[63202]: DEBUG oslo_vmware.api [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385878, 'name': ReconfigVM_Task, 'duration_secs': 0.174398} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.200090] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294210', 'volume_id': '81b406ab-e721-4271-b298-75cded690f20', 'name': 'volume-81b406ab-e721-4271-b298-75cded690f20', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '8437f856-a707-49c5-b8eb-5a22cdb990f8', 'attached_at': '', 'detached_at': '', 'volume_id': '81b406ab-e721-4271-b298-75cded690f20', 'serial': '81b406ab-e721-4271-b298-75cded690f20'} {{(pid=63202) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1067.200599] env[63202]: DEBUG oslo_concurrency.lockutils [req-cffd8f7d-0cc4-4de1-b7d7-e1eb9075c5f6 req-e74fa45c-6710-4775-9019-b098955ca441 service nova] Releasing lock "refresh_cache-1123a12e-5218-415d-b286-2f005fe57b29" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1067.200894] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d6592eab-ca6c-408b-8d23-78473bc3f428 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.207230] env[63202]: DEBUG oslo_vmware.api [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1067.207230] env[63202]: value = "task-1385880" [ 1067.207230] env[63202]: _type = "Task" [ 1067.207230] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.217550] env[63202]: DEBUG oslo_vmware.api [None req-a31960a5-3c43-4c5a-a5a0-a9585adc3229 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385879, 'name': SuspendVM_Task} progress is 62%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.220535] env[63202]: DEBUG oslo_vmware.api [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385880, 'name': Rename_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.283110] env[63202]: DEBUG nova.scheduler.client.report [None req-2ef4997b-172c-4d17-b6cd-0f645eb38a09 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1067.568318] env[63202]: DEBUG oslo_vmware.api [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52877c3a-a745-d108-abc7-b75a187ba919, 'name': SearchDatastore_Task, 'duration_secs': 0.009596} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.568318] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1067.568318] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1067.568318] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1067.568613] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.568613] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1067.568613] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ff27d7b4-8877-4d8c-b318-143994edae53 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.576712] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1067.576894] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1067.577652] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-061f11ed-6664-40ef-8505-5bb40b7584b2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.581095] env[63202]: DEBUG nova.network.neutron [req-c23354e6-f223-4abd-b570-433e361ad6f0 req-6b328960-156d-481d-8179-910bcaabba1e service nova] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Updated VIF entry in instance network info cache for port aaf1e261-e958-4575-a0c8-b9959bf1bd4d. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1067.581577] env[63202]: DEBUG nova.network.neutron [req-c23354e6-f223-4abd-b570-433e361ad6f0 req-6b328960-156d-481d-8179-910bcaabba1e service nova] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Updating instance_info_cache with network_info: [{"id": "aaf1e261-e958-4575-a0c8-b9959bf1bd4d", "address": "fa:16:3e:96:4d:5a", "network": {"id": "b906037c-ca02-45b3-83f0-620a7a89656c", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-877611281-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b0fe5199b8b940e0b23c49026fdcc685", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48512b02-ad5c-4105-ba7d-fd4775acf8e1", "external-id": "nsx-vlan-transportzone-516", "segmentation_id": 516, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaaf1e261-e9", "ovs_interfaceid": "aaf1e261-e958-4575-a0c8-b9959bf1bd4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1067.584028] env[63202]: DEBUG oslo_vmware.api [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Waiting for the task: (returnval){ [ 1067.584028] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]521c8d5b-ad16-765f-6195-1a626c3336e7" [ 1067.584028] env[63202]: _type = "Task" [ 1067.584028] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.594521] env[63202]: DEBUG oslo_vmware.api [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]521c8d5b-ad16-765f-6195-1a626c3336e7, 'name': SearchDatastore_Task, 'duration_secs': 0.009068} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.595749] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd2716a7-559f-4eb4-aa0a-484642e06fce {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.601339] env[63202]: DEBUG oslo_vmware.api [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Waiting for the task: (returnval){ [ 1067.601339] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]525b9730-bd69-1fb7-2934-12c8a59f974e" [ 1067.601339] env[63202]: _type = "Task" [ 1067.601339] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.609489] env[63202]: DEBUG oslo_vmware.api [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]525b9730-bd69-1fb7-2934-12c8a59f974e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.621533] env[63202]: DEBUG nova.network.neutron [-] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1067.719773] env[63202]: DEBUG oslo_vmware.api [None req-a31960a5-3c43-4c5a-a5a0-a9585adc3229 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385879, 'name': SuspendVM_Task, 'duration_secs': 0.606799} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.722764] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a31960a5-3c43-4c5a-a5a0-a9585adc3229 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Suspended the VM {{(pid=63202) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1067.722961] env[63202]: DEBUG nova.compute.manager [None req-a31960a5-3c43-4c5a-a5a0-a9585adc3229 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1067.723251] env[63202]: DEBUG oslo_vmware.api [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385880, 'name': Rename_Task, 'duration_secs': 0.146153} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.723924] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d1b195c-fde2-48d3-8573-47a9db3b7fb9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.726282] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1067.726514] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bbd7010e-bb63-4e81-920e-b7de17b8ddc6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.734149] env[63202]: DEBUG oslo_vmware.api [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1067.734149] env[63202]: value = "task-1385881" [ 1067.734149] env[63202]: _type = "Task" [ 1067.734149] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.741695] env[63202]: DEBUG oslo_vmware.api [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385881, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.788692] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2ef4997b-172c-4d17-b6cd-0f645eb38a09 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.708s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.791020] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.314s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.792674] env[63202]: INFO nova.compute.claims [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1067.813623] env[63202]: INFO nova.scheduler.client.report [None req-2ef4997b-172c-4d17-b6cd-0f645eb38a09 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Deleted allocations for instance 7437595c-fa35-483e-95f3-b75405b6bd13 [ 1068.086064] env[63202]: DEBUG oslo_concurrency.lockutils [req-c23354e6-f223-4abd-b570-433e361ad6f0 req-6b328960-156d-481d-8179-910bcaabba1e service nova] Releasing lock "refresh_cache-6e172a2a-1918-430c-84ce-005fef97bbf1" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1068.113333] env[63202]: DEBUG oslo_vmware.api [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]525b9730-bd69-1fb7-2934-12c8a59f974e, 'name': SearchDatastore_Task, 'duration_secs': 0.008743} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.113608] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1068.113871] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore2] 6e172a2a-1918-430c-84ce-005fef97bbf1/6e172a2a-1918-430c-84ce-005fef97bbf1.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1068.114464] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fac83f71-f731-4033-bb27-7cc018b4bdfb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.121248] env[63202]: DEBUG oslo_vmware.api [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Waiting for the task: (returnval){ [ 1068.121248] env[63202]: value = "task-1385882" [ 1068.121248] env[63202]: _type = "Task" [ 1068.121248] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.124665] env[63202]: INFO nova.compute.manager [-] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Took 1.27 seconds to deallocate network for instance. [ 1068.132580] env[63202]: DEBUG oslo_vmware.api [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Task: {'id': task-1385882, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.250033] env[63202]: DEBUG oslo_vmware.api [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385881, 'name': PowerOnVM_Task, 'duration_secs': 0.484519} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.250205] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1068.320497] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2ef4997b-172c-4d17-b6cd-0f645eb38a09 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "7437595c-fa35-483e-95f3-b75405b6bd13" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.393s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.368831] env[63202]: DEBUG nova.compute.manager [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1068.369348] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c6c067a-d671-4354-8742-15df5cb10362 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.633260] env[63202]: DEBUG oslo_vmware.api [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Task: {'id': task-1385882, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.636793] env[63202]: DEBUG oslo_concurrency.lockutils [None req-fe0ebb9d-359a-44f5-9fe9-5ec465a84793 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.841889] env[63202]: DEBUG nova.compute.manager [req-12827b4e-58c1-40e1-8a91-13ce19d4a3c3 req-437d1317-8402-4100-8327-4f263775395a service nova] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Received event network-vif-deleted-b78164a7-4413-4943-b4fd-0a55d84f87fe {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1068.891205] env[63202]: DEBUG oslo_concurrency.lockutils [None req-caff465d-ca2f-4313-98a5-fd3e40a40baa tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lock "8437f856-a707-49c5-b8eb-5a22cdb990f8" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 29.227s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.896996] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f81bc13-7b5c-49ca-a994-009765cb329c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.903447] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a3f947f-d489-4f2f-bced-7cf2b6d334c1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.933706] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-487d7ebf-c02c-43f7-8665-0509a3b9d310 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.941063] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-190f5deb-f781-40e4-8cce-c552c2102417 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.954378] env[63202]: DEBUG nova.compute.provider_tree [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1069.131249] env[63202]: DEBUG oslo_vmware.api [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Task: {'id': task-1385882, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.733762} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.131587] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore2] 6e172a2a-1918-430c-84ce-005fef97bbf1/6e172a2a-1918-430c-84ce-005fef97bbf1.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1069.131783] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1069.132063] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-93623c7b-51fd-422a-a13c-3a0ed038a5da {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.140777] env[63202]: DEBUG oslo_vmware.api [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Waiting for the task: (returnval){ [ 1069.140777] env[63202]: value = "task-1385883" [ 1069.140777] env[63202]: _type = "Task" [ 1069.140777] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.147976] env[63202]: DEBUG oslo_vmware.api [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Task: {'id': task-1385883, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.193870] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5026b813-0997-45b3-bd8e-32c84c1849c2 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "439ee94c-04f2-45d3-a486-81a216c8db4e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.194145] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5026b813-0997-45b3-bd8e-32c84c1849c2 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "439ee94c-04f2-45d3-a486-81a216c8db4e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.194369] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5026b813-0997-45b3-bd8e-32c84c1849c2 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "439ee94c-04f2-45d3-a486-81a216c8db4e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.194560] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5026b813-0997-45b3-bd8e-32c84c1849c2 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "439ee94c-04f2-45d3-a486-81a216c8db4e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.194734] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5026b813-0997-45b3-bd8e-32c84c1849c2 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "439ee94c-04f2-45d3-a486-81a216c8db4e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.197457] env[63202]: INFO nova.compute.manager [None req-5026b813-0997-45b3-bd8e-32c84c1849c2 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Terminating instance [ 1069.199027] env[63202]: DEBUG nova.compute.manager [None req-5026b813-0997-45b3-bd8e-32c84c1849c2 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1069.199139] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-5026b813-0997-45b3-bd8e-32c84c1849c2 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1069.200011] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22332c2f-6e89-4804-bdec-4cddac90ae40 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.207862] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-5026b813-0997-45b3-bd8e-32c84c1849c2 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1069.208164] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8c63db03-de14-48b1-ac46-802b51595bd2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.254098] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1069.254286] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Starting heal instance info cache {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1069.254407] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Rebuilding the list of instances to heal {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1069.285318] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-5026b813-0997-45b3-bd8e-32c84c1849c2 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1069.285934] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-5026b813-0997-45b3-bd8e-32c84c1849c2 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Deleting contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1069.285934] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-5026b813-0997-45b3-bd8e-32c84c1849c2 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Deleting the datastore file [datastore1] 439ee94c-04f2-45d3-a486-81a216c8db4e {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1069.286130] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-51a5797c-4077-45d5-992e-91ae4cb51a2b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.293378] env[63202]: DEBUG oslo_vmware.api [None req-5026b813-0997-45b3-bd8e-32c84c1849c2 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 1069.293378] env[63202]: value = "task-1385885" [ 1069.293378] env[63202]: _type = "Task" [ 1069.293378] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.301076] env[63202]: DEBUG oslo_vmware.api [None req-5026b813-0997-45b3-bd8e-32c84c1849c2 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385885, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.457323] env[63202]: DEBUG nova.scheduler.client.report [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1069.651056] env[63202]: DEBUG oslo_vmware.api [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Task: {'id': task-1385883, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086907} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.651056] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1069.651684] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f58c362e-27f7-4ca1-84e3-e9bb9b04a956 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.673072] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] 6e172a2a-1918-430c-84ce-005fef97bbf1/6e172a2a-1918-430c-84ce-005fef97bbf1.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1069.673332] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e0db4334-b511-4f22-b749-d1cc5c0047d0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.691709] env[63202]: DEBUG oslo_vmware.api [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Waiting for the task: (returnval){ [ 1069.691709] env[63202]: value = "task-1385886" [ 1069.691709] env[63202]: _type = "Task" [ 1069.691709] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.699190] env[63202]: DEBUG oslo_vmware.api [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Task: {'id': task-1385886, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.761754] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Skipping network cache update for instance because it is being deleted. {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 1069.761943] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Skipping network cache update for instance because it is Building. {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 1069.762125] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Skipping network cache update for instance because it is Building. {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 1069.793539] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Acquiring lock "refresh_cache-8437f856-a707-49c5-b8eb-5a22cdb990f8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1069.793684] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Acquired lock "refresh_cache-8437f856-a707-49c5-b8eb-5a22cdb990f8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1069.794014] env[63202]: DEBUG nova.network.neutron [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Forcefully refreshing network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1069.794014] env[63202]: DEBUG nova.objects.instance [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lazy-loading 'info_cache' on Instance uuid 8437f856-a707-49c5-b8eb-5a22cdb990f8 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1069.806051] env[63202]: DEBUG oslo_vmware.api [None req-5026b813-0997-45b3-bd8e-32c84c1849c2 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385885, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.161676} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.807050] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-5026b813-0997-45b3-bd8e-32c84c1849c2 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1069.807263] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-5026b813-0997-45b3-bd8e-32c84c1849c2 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Deleted contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1069.807572] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-5026b813-0997-45b3-bd8e-32c84c1849c2 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1069.807680] env[63202]: INFO nova.compute.manager [None req-5026b813-0997-45b3-bd8e-32c84c1849c2 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Took 0.61 seconds to destroy the instance on the hypervisor. [ 1069.807986] env[63202]: DEBUG oslo.service.loopingcall [None req-5026b813-0997-45b3-bd8e-32c84c1849c2 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1069.808520] env[63202]: DEBUG nova.compute.manager [-] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1069.808635] env[63202]: DEBUG nova.network.neutron [-] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1069.963218] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.171s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.963218] env[63202]: DEBUG nova.compute.manager [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1069.965893] env[63202]: DEBUG oslo_concurrency.lockutils [None req-fe0ebb9d-359a-44f5-9fe9-5ec465a84793 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.330s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.965893] env[63202]: DEBUG nova.objects.instance [None req-fe0ebb9d-359a-44f5-9fe9-5ec465a84793 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Lazy-loading 'resources' on Instance uuid 1123a12e-5218-415d-b286-2f005fe57b29 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1070.201710] env[63202]: DEBUG oslo_vmware.api [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Task: {'id': task-1385886, 'name': ReconfigVM_Task, 'duration_secs': 0.264751} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.202049] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Reconfigured VM instance instance-00000065 to attach disk [datastore2] 6e172a2a-1918-430c-84ce-005fef97bbf1/6e172a2a-1918-430c-84ce-005fef97bbf1.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1070.202671] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5fa65731-46fc-45cf-86fa-9cb7cd60eb78 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.208555] env[63202]: DEBUG oslo_vmware.api [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Waiting for the task: (returnval){ [ 1070.208555] env[63202]: value = "task-1385887" [ 1070.208555] env[63202]: _type = "Task" [ 1070.208555] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.216829] env[63202]: DEBUG oslo_vmware.api [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Task: {'id': task-1385887, 'name': Rename_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.404725] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquiring lock "214c63b6-35e3-46e1-a9de-f3cea251c3c9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.404971] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "214c63b6-35e3-46e1-a9de-f3cea251c3c9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.468502] env[63202]: DEBUG nova.compute.utils [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1070.472506] env[63202]: DEBUG nova.compute.manager [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1070.472695] env[63202]: DEBUG nova.network.neutron [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1070.519053] env[63202]: DEBUG nova.policy [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '47a40b09812d4be68bb478771514dfa3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '433b1d11832147b886bed7a3a4952768', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 1070.559202] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df5a48cd-ad99-45c6-8217-3c3415e93396 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.563675] env[63202]: DEBUG nova.network.neutron [-] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1070.567614] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76650377-a28f-4cff-b76e-da3ee8c2114a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.598072] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd01a0ae-ec8a-49db-adbf-590b17f3336f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.606633] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8822d61e-694e-4cb3-8c27-b466ace2c1bb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.621192] env[63202]: DEBUG nova.compute.provider_tree [None req-fe0ebb9d-359a-44f5-9fe9-5ec465a84793 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1070.720218] env[63202]: DEBUG oslo_vmware.api [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Task: {'id': task-1385887, 'name': Rename_Task, 'duration_secs': 0.131315} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.720490] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1070.720734] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9b38a15f-41c1-4790-9744-6696e21ed4b7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.727325] env[63202]: DEBUG oslo_vmware.api [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Waiting for the task: (returnval){ [ 1070.727325] env[63202]: value = "task-1385888" [ 1070.727325] env[63202]: _type = "Task" [ 1070.727325] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.734855] env[63202]: DEBUG oslo_vmware.api [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Task: {'id': task-1385888, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.786149] env[63202]: DEBUG nova.network.neutron [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Successfully created port: 55537461-97ca-40a4-8086-69473669d5b6 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1070.890953] env[63202]: DEBUG nova.compute.manager [req-0e4a7bb3-24f6-4235-b937-c99fd1f6a7b0 req-7244d45a-2043-41d8-85c1-e05ac8d499d9 service nova] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Received event network-vif-deleted-a8eb92d1-368f-42cc-83d2-d516202801a3 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1070.906937] env[63202]: DEBUG nova.compute.manager [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1070.973338] env[63202]: DEBUG nova.compute.manager [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1071.067399] env[63202]: INFO nova.compute.manager [-] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Took 1.26 seconds to deallocate network for instance. [ 1071.124161] env[63202]: DEBUG nova.scheduler.client.report [None req-fe0ebb9d-359a-44f5-9fe9-5ec465a84793 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1071.236789] env[63202]: DEBUG oslo_vmware.api [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Task: {'id': task-1385888, 'name': PowerOnVM_Task, 'duration_secs': 0.439735} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.237113] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1071.237240] env[63202]: INFO nova.compute.manager [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Took 8.01 seconds to spawn the instance on the hypervisor. [ 1071.237386] env[63202]: DEBUG nova.compute.manager [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1071.238146] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75fddbcb-a81a-44c6-85e1-f1caf6113f79 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.429706] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1071.534024] env[63202]: DEBUG nova.network.neutron [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Updating instance_info_cache with network_info: [{"id": "07cc827a-876a-4432-a5fc-9ba8920dc5f9", "address": "fa:16:3e:71:e6:9f", "network": {"id": "06ab5813-9ad9-4021-9bdb-f2f02af8d73f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1714653503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.175", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b2de35030a9484094e964ffc30a822d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3093647a-bab7-4562-ada0-428725e8c0fc", "external-id": "nsx-vlan-transportzone-660", "segmentation_id": 660, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap07cc827a-87", "ovs_interfaceid": "07cc827a-876a-4432-a5fc-9ba8920dc5f9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1071.573686] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5026b813-0997-45b3-bd8e-32c84c1849c2 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1071.629243] env[63202]: DEBUG oslo_concurrency.lockutils [None req-fe0ebb9d-359a-44f5-9fe9-5ec465a84793 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.664s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.632046] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.202s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1071.633020] env[63202]: INFO nova.compute.claims [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1071.647803] env[63202]: INFO nova.scheduler.client.report [None req-fe0ebb9d-359a-44f5-9fe9-5ec465a84793 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Deleted allocations for instance 1123a12e-5218-415d-b286-2f005fe57b29 [ 1071.755568] env[63202]: INFO nova.compute.manager [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Took 17.54 seconds to build instance. [ 1071.983454] env[63202]: DEBUG nova.compute.manager [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1072.007525] env[63202]: DEBUG nova.virt.hardware [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1072.007820] env[63202]: DEBUG nova.virt.hardware [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1072.007986] env[63202]: DEBUG nova.virt.hardware [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1072.008176] env[63202]: DEBUG nova.virt.hardware [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1072.008327] env[63202]: DEBUG nova.virt.hardware [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1072.008479] env[63202]: DEBUG nova.virt.hardware [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1072.008688] env[63202]: DEBUG nova.virt.hardware [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1072.008852] env[63202]: DEBUG nova.virt.hardware [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1072.009091] env[63202]: DEBUG nova.virt.hardware [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1072.009302] env[63202]: DEBUG nova.virt.hardware [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1072.009543] env[63202]: DEBUG nova.virt.hardware [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1072.010463] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d766e4dc-f1e7-46c4-8f76-4894ac0dec11 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.018372] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68837a30-b806-4016-a7e5-14075180ceb2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.036051] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Releasing lock "refresh_cache-8437f856-a707-49c5-b8eb-5a22cdb990f8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1072.036243] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Updated the network info_cache for instance {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1072.155748] env[63202]: DEBUG oslo_concurrency.lockutils [None req-fe0ebb9d-359a-44f5-9fe9-5ec465a84793 tempest-ServerRescueTestJSONUnderV235-1601333119 tempest-ServerRescueTestJSONUnderV235-1601333119-project-member] Lock "1123a12e-5218-415d-b286-2f005fe57b29" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.473s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.257216] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89b799d0-7d04-4df2-b51d-3549c54814bb tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Lock "6e172a2a-1918-430c-84ce-005fef97bbf1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.054s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.393401] env[63202]: DEBUG nova.network.neutron [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Successfully updated port: 55537461-97ca-40a4-8086-69473669d5b6 {{(pid=63202) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1072.566486] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c3f9b308-d144-49dc-8dcc-b99be23ef6ad tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Acquiring lock "6e172a2a-1918-430c-84ce-005fef97bbf1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.566762] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c3f9b308-d144-49dc-8dcc-b99be23ef6ad tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Lock "6e172a2a-1918-430c-84ce-005fef97bbf1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.566973] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c3f9b308-d144-49dc-8dcc-b99be23ef6ad tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Acquiring lock "6e172a2a-1918-430c-84ce-005fef97bbf1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.567196] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c3f9b308-d144-49dc-8dcc-b99be23ef6ad tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Lock "6e172a2a-1918-430c-84ce-005fef97bbf1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.567379] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c3f9b308-d144-49dc-8dcc-b99be23ef6ad tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Lock "6e172a2a-1918-430c-84ce-005fef97bbf1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.570957] env[63202]: INFO nova.compute.manager [None req-c3f9b308-d144-49dc-8dcc-b99be23ef6ad tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Terminating instance [ 1072.571759] env[63202]: DEBUG nova.compute.manager [None req-c3f9b308-d144-49dc-8dcc-b99be23ef6ad tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1072.571952] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c3f9b308-d144-49dc-8dcc-b99be23ef6ad tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1072.572877] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e8a92bd-07b6-488f-bdbf-d9083f761c21 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.581145] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3f9b308-d144-49dc-8dcc-b99be23ef6ad tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1072.581638] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-39a11f69-76a4-40b6-8989-95cfeb85cf5e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.587842] env[63202]: DEBUG oslo_vmware.api [None req-c3f9b308-d144-49dc-8dcc-b99be23ef6ad tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Waiting for the task: (returnval){ [ 1072.587842] env[63202]: value = "task-1385889" [ 1072.587842] env[63202]: _type = "Task" [ 1072.587842] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.596407] env[63202]: DEBUG oslo_vmware.api [None req-c3f9b308-d144-49dc-8dcc-b99be23ef6ad tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Task: {'id': task-1385889, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.738633] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c307ed64-4e64-4faf-a246-1ac946d6c390 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.746065] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9f4a97d-0ab4-4a32-91c9-82378441d255 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.777041] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2d4625c-aa0d-40f1-9687-d19b400a8862 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.784750] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-142a77d8-24bd-40c0-bd9d-be29f9855bec {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.802900] env[63202]: DEBUG nova.compute.provider_tree [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1072.903694] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Acquiring lock "refresh_cache-9646810e-06d2-4de0-8f0c-e8719541da53" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1072.903694] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Acquired lock "refresh_cache-9646810e-06d2-4de0-8f0c-e8719541da53" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1072.903694] env[63202]: DEBUG nova.network.neutron [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1072.917378] env[63202]: DEBUG nova.compute.manager [req-baeb4113-1012-4b65-82bb-7c15bfdd379d req-c537ee0f-6d94-4fc2-8581-d760844fc1aa service nova] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Received event network-vif-plugged-55537461-97ca-40a4-8086-69473669d5b6 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1072.917598] env[63202]: DEBUG oslo_concurrency.lockutils [req-baeb4113-1012-4b65-82bb-7c15bfdd379d req-c537ee0f-6d94-4fc2-8581-d760844fc1aa service nova] Acquiring lock "9646810e-06d2-4de0-8f0c-e8719541da53-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.917805] env[63202]: DEBUG oslo_concurrency.lockutils [req-baeb4113-1012-4b65-82bb-7c15bfdd379d req-c537ee0f-6d94-4fc2-8581-d760844fc1aa service nova] Lock "9646810e-06d2-4de0-8f0c-e8719541da53-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.917971] env[63202]: DEBUG oslo_concurrency.lockutils [req-baeb4113-1012-4b65-82bb-7c15bfdd379d req-c537ee0f-6d94-4fc2-8581-d760844fc1aa service nova] Lock "9646810e-06d2-4de0-8f0c-e8719541da53-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.918163] env[63202]: DEBUG nova.compute.manager [req-baeb4113-1012-4b65-82bb-7c15bfdd379d req-c537ee0f-6d94-4fc2-8581-d760844fc1aa service nova] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] No waiting events found dispatching network-vif-plugged-55537461-97ca-40a4-8086-69473669d5b6 {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1072.918332] env[63202]: WARNING nova.compute.manager [req-baeb4113-1012-4b65-82bb-7c15bfdd379d req-c537ee0f-6d94-4fc2-8581-d760844fc1aa service nova] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Received unexpected event network-vif-plugged-55537461-97ca-40a4-8086-69473669d5b6 for instance with vm_state building and task_state spawning. [ 1072.918494] env[63202]: DEBUG nova.compute.manager [req-baeb4113-1012-4b65-82bb-7c15bfdd379d req-c537ee0f-6d94-4fc2-8581-d760844fc1aa service nova] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Received event network-changed-55537461-97ca-40a4-8086-69473669d5b6 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1072.918649] env[63202]: DEBUG nova.compute.manager [req-baeb4113-1012-4b65-82bb-7c15bfdd379d req-c537ee0f-6d94-4fc2-8581-d760844fc1aa service nova] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Refreshing instance network info cache due to event network-changed-55537461-97ca-40a4-8086-69473669d5b6. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1072.918816] env[63202]: DEBUG oslo_concurrency.lockutils [req-baeb4113-1012-4b65-82bb-7c15bfdd379d req-c537ee0f-6d94-4fc2-8581-d760844fc1aa service nova] Acquiring lock "refresh_cache-9646810e-06d2-4de0-8f0c-e8719541da53" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1073.097838] env[63202]: DEBUG oslo_vmware.api [None req-c3f9b308-d144-49dc-8dcc-b99be23ef6ad tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Task: {'id': task-1385889, 'name': PowerOffVM_Task, 'duration_secs': 0.207537} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.098145] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3f9b308-d144-49dc-8dcc-b99be23ef6ad tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1073.098305] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c3f9b308-d144-49dc-8dcc-b99be23ef6ad tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1073.098562] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-855d14c7-df03-423e-a846-b66accb26498 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.182188] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c3f9b308-d144-49dc-8dcc-b99be23ef6ad tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1073.182435] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c3f9b308-d144-49dc-8dcc-b99be23ef6ad tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Deleting contents of the VM from datastore datastore2 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1073.182653] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3f9b308-d144-49dc-8dcc-b99be23ef6ad tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Deleting the datastore file [datastore2] 6e172a2a-1918-430c-84ce-005fef97bbf1 {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1073.182920] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c4ba44ec-0e17-4e22-a673-e185a6b79d95 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.189777] env[63202]: DEBUG oslo_vmware.api [None req-c3f9b308-d144-49dc-8dcc-b99be23ef6ad tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Waiting for the task: (returnval){ [ 1073.189777] env[63202]: value = "task-1385891" [ 1073.189777] env[63202]: _type = "Task" [ 1073.189777] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.197623] env[63202]: DEBUG oslo_vmware.api [None req-c3f9b308-d144-49dc-8dcc-b99be23ef6ad tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Task: {'id': task-1385891, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.306098] env[63202]: DEBUG nova.scheduler.client.report [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1073.431889] env[63202]: DEBUG nova.network.neutron [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1073.579315] env[63202]: DEBUG nova.network.neutron [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Updating instance_info_cache with network_info: [{"id": "55537461-97ca-40a4-8086-69473669d5b6", "address": "fa:16:3e:cc:19:9e", "network": {"id": "f105b723-b2a9-401d-a936-e31b4b6609f6", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-298435423-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "433b1d11832147b886bed7a3a4952768", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e4c8c8fd-baca-4e60-97dc-ff0418d63215", "external-id": "nsx-vlan-transportzone-178", "segmentation_id": 178, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap55537461-97", "ovs_interfaceid": "55537461-97ca-40a4-8086-69473669d5b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1073.700062] env[63202]: DEBUG oslo_vmware.api [None req-c3f9b308-d144-49dc-8dcc-b99be23ef6ad tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Task: {'id': task-1385891, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.158722} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.700337] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3f9b308-d144-49dc-8dcc-b99be23ef6ad tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1073.700529] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c3f9b308-d144-49dc-8dcc-b99be23ef6ad tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Deleted contents of the VM from datastore datastore2 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1073.700714] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c3f9b308-d144-49dc-8dcc-b99be23ef6ad tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1073.700891] env[63202]: INFO nova.compute.manager [None req-c3f9b308-d144-49dc-8dcc-b99be23ef6ad tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1073.701167] env[63202]: DEBUG oslo.service.loopingcall [None req-c3f9b308-d144-49dc-8dcc-b99be23ef6ad tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1073.701367] env[63202]: DEBUG nova.compute.manager [-] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1073.701460] env[63202]: DEBUG nova.network.neutron [-] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1073.812696] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.181s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.813198] env[63202]: DEBUG nova.compute.manager [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1073.815872] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5026b813-0997-45b3-bd8e-32c84c1849c2 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.242s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.816101] env[63202]: DEBUG nova.objects.instance [None req-5026b813-0997-45b3-bd8e-32c84c1849c2 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lazy-loading 'resources' on Instance uuid 439ee94c-04f2-45d3-a486-81a216c8db4e {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1074.082289] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Releasing lock "refresh_cache-9646810e-06d2-4de0-8f0c-e8719541da53" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1074.082577] env[63202]: DEBUG nova.compute.manager [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Instance network_info: |[{"id": "55537461-97ca-40a4-8086-69473669d5b6", "address": "fa:16:3e:cc:19:9e", "network": {"id": "f105b723-b2a9-401d-a936-e31b4b6609f6", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-298435423-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "433b1d11832147b886bed7a3a4952768", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e4c8c8fd-baca-4e60-97dc-ff0418d63215", "external-id": "nsx-vlan-transportzone-178", "segmentation_id": 178, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap55537461-97", "ovs_interfaceid": "55537461-97ca-40a4-8086-69473669d5b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1074.082927] env[63202]: DEBUG oslo_concurrency.lockutils [req-baeb4113-1012-4b65-82bb-7c15bfdd379d req-c537ee0f-6d94-4fc2-8581-d760844fc1aa service nova] Acquired lock "refresh_cache-9646810e-06d2-4de0-8f0c-e8719541da53" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1074.083127] env[63202]: DEBUG nova.network.neutron [req-baeb4113-1012-4b65-82bb-7c15bfdd379d req-c537ee0f-6d94-4fc2-8581-d760844fc1aa service nova] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Refreshing network info cache for port 55537461-97ca-40a4-8086-69473669d5b6 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1074.084280] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cc:19:9e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e4c8c8fd-baca-4e60-97dc-ff0418d63215', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '55537461-97ca-40a4-8086-69473669d5b6', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1074.091780] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Creating folder: Project (433b1d11832147b886bed7a3a4952768). Parent ref: group-v294090. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1074.095091] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-14df7a51-c522-4da6-8a38-1a85e746c192 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.166584] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Created folder: Project (433b1d11832147b886bed7a3a4952768) in parent group-v294090. [ 1074.166831] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Creating folder: Instances. Parent ref: group-v294221. {{(pid=63202) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1074.168831] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9c5d43de-3be6-4ba8-90a8-dc9135238a27 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.177492] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Created folder: Instances in parent group-v294221. [ 1074.178285] env[63202]: DEBUG oslo.service.loopingcall [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1074.178285] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1074.178285] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c41e103f-2929-4048-8032-a9413a403d93 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.199603] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1074.199603] env[63202]: value = "task-1385894" [ 1074.199603] env[63202]: _type = "Task" [ 1074.199603] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.206918] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385894, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.295660] env[63202]: DEBUG nova.network.neutron [req-baeb4113-1012-4b65-82bb-7c15bfdd379d req-c537ee0f-6d94-4fc2-8581-d760844fc1aa service nova] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Updated VIF entry in instance network info cache for port 55537461-97ca-40a4-8086-69473669d5b6. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1074.296083] env[63202]: DEBUG nova.network.neutron [req-baeb4113-1012-4b65-82bb-7c15bfdd379d req-c537ee0f-6d94-4fc2-8581-d760844fc1aa service nova] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Updating instance_info_cache with network_info: [{"id": "55537461-97ca-40a4-8086-69473669d5b6", "address": "fa:16:3e:cc:19:9e", "network": {"id": "f105b723-b2a9-401d-a936-e31b4b6609f6", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-298435423-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "433b1d11832147b886bed7a3a4952768", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e4c8c8fd-baca-4e60-97dc-ff0418d63215", "external-id": "nsx-vlan-transportzone-178", "segmentation_id": 178, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap55537461-97", "ovs_interfaceid": "55537461-97ca-40a4-8086-69473669d5b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1074.318675] env[63202]: DEBUG nova.compute.utils [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1074.323440] env[63202]: DEBUG nova.compute.manager [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1074.323602] env[63202]: DEBUG nova.network.neutron [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1074.376162] env[63202]: DEBUG nova.policy [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '63a29c785cf240a7b6418dadc7119574', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f7d28fea097541adadf0839940568409', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 1074.417925] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d189d07-87b9-4b07-a26e-43fdf05761c4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.425282] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b561aa2c-fd28-47a5-a3e6-412439dca476 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.455682] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47348ec5-eaf3-44ff-91df-95770738c343 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.463624] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e30271d-7f85-4f32-8d91-6edb8fbebc8f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.467876] env[63202]: DEBUG nova.network.neutron [-] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1074.478831] env[63202]: DEBUG nova.compute.provider_tree [None req-5026b813-0997-45b3-bd8e-32c84c1849c2 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1074.657993] env[63202]: DEBUG nova.network.neutron [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Successfully created port: 76e223aa-1b1a-4d12-a8a6-89bb986656b5 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1074.708918] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385894, 'name': CreateVM_Task, 'duration_secs': 0.320401} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.709124] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1074.709836] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1074.710013] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1074.710349] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1074.710611] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4751cd96-e73b-4c9c-a92d-87c83ad7d91c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.715389] env[63202]: DEBUG oslo_vmware.api [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Waiting for the task: (returnval){ [ 1074.715389] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52a7cc47-057b-246c-c5b4-503835dfbfcf" [ 1074.715389] env[63202]: _type = "Task" [ 1074.715389] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.722979] env[63202]: DEBUG oslo_vmware.api [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52a7cc47-057b-246c-c5b4-503835dfbfcf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.798533] env[63202]: DEBUG oslo_concurrency.lockutils [req-baeb4113-1012-4b65-82bb-7c15bfdd379d req-c537ee0f-6d94-4fc2-8581-d760844fc1aa service nova] Releasing lock "refresh_cache-9646810e-06d2-4de0-8f0c-e8719541da53" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1074.824458] env[63202]: DEBUG nova.compute.manager [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1074.950519] env[63202]: DEBUG nova.compute.manager [req-c1113917-c86b-424d-a250-b178e9c60a7e req-c5b54659-c568-4d3d-af01-c7f64d8b028a service nova] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Received event network-vif-deleted-aaf1e261-e958-4575-a0c8-b9959bf1bd4d {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1074.970928] env[63202]: INFO nova.compute.manager [-] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Took 1.27 seconds to deallocate network for instance. [ 1074.981691] env[63202]: DEBUG nova.scheduler.client.report [None req-5026b813-0997-45b3-bd8e-32c84c1849c2 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1075.225387] env[63202]: DEBUG oslo_vmware.api [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52a7cc47-057b-246c-c5b4-503835dfbfcf, 'name': SearchDatastore_Task, 'duration_secs': 0.01174} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.225698] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1075.226156] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1075.226214] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1075.226432] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1075.226499] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1075.226754] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fa469dbe-b95e-450e-a3ec-828deac5f284 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.235843] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1075.236066] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1075.236816] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e96e117e-e5b9-4483-b6a0-1e8b8517a8d1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.242447] env[63202]: DEBUG oslo_vmware.api [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Waiting for the task: (returnval){ [ 1075.242447] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52dbf5d6-7dbd-8dc9-7c44-bbae146fcafd" [ 1075.242447] env[63202]: _type = "Task" [ 1075.242447] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.251104] env[63202]: DEBUG oslo_vmware.api [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52dbf5d6-7dbd-8dc9-7c44-bbae146fcafd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.478818] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c3f9b308-d144-49dc-8dcc-b99be23ef6ad tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.486931] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5026b813-0997-45b3-bd8e-32c84c1849c2 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.671s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.489759] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c3f9b308-d144-49dc-8dcc-b99be23ef6ad tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.011s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.490128] env[63202]: DEBUG nova.objects.instance [None req-c3f9b308-d144-49dc-8dcc-b99be23ef6ad tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Lazy-loading 'resources' on Instance uuid 6e172a2a-1918-430c-84ce-005fef97bbf1 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1075.516230] env[63202]: INFO nova.scheduler.client.report [None req-5026b813-0997-45b3-bd8e-32c84c1849c2 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Deleted allocations for instance 439ee94c-04f2-45d3-a486-81a216c8db4e [ 1075.753666] env[63202]: DEBUG oslo_vmware.api [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52dbf5d6-7dbd-8dc9-7c44-bbae146fcafd, 'name': SearchDatastore_Task, 'duration_secs': 0.00935} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.754607] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-662b26eb-dd5f-428e-b6aa-db2ce3ff0a61 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.760164] env[63202]: DEBUG oslo_vmware.api [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Waiting for the task: (returnval){ [ 1075.760164] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52bdb80c-86aa-bd1d-8ee0-5879e607f1d5" [ 1075.760164] env[63202]: _type = "Task" [ 1075.760164] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.768306] env[63202]: DEBUG oslo_vmware.api [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52bdb80c-86aa-bd1d-8ee0-5879e607f1d5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.835770] env[63202]: DEBUG nova.compute.manager [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1075.864387] env[63202]: DEBUG nova.virt.hardware [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1075.864713] env[63202]: DEBUG nova.virt.hardware [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1075.864909] env[63202]: DEBUG nova.virt.hardware [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1075.865118] env[63202]: DEBUG nova.virt.hardware [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1075.865312] env[63202]: DEBUG nova.virt.hardware [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1075.865510] env[63202]: DEBUG nova.virt.hardware [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1075.865746] env[63202]: DEBUG nova.virt.hardware [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1075.865967] env[63202]: DEBUG nova.virt.hardware [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1075.866202] env[63202]: DEBUG nova.virt.hardware [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1075.866384] env[63202]: DEBUG nova.virt.hardware [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1075.866621] env[63202]: DEBUG nova.virt.hardware [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1075.867647] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3ad63e3-3faa-4fee-b85f-64c71340431f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.876406] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d16a294f-7f5a-4d18-a2f7-3ac4f8dd85bf {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.025212] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5026b813-0997-45b3-bd8e-32c84c1849c2 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "439ee94c-04f2-45d3-a486-81a216c8db4e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.831s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1076.088174] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd36c019-93f2-4140-8ac9-955a9ef1f1c0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.096276] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c9af9b1-f4cc-4a19-a017-80b85845ef3d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.128270] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de3e16dd-81cb-49f9-a5cb-3077ee589e89 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.136284] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b95a16e0-0882-45bd-a356-7fa05f9d8ada {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.150246] env[63202]: DEBUG nova.compute.provider_tree [None req-c3f9b308-d144-49dc-8dcc-b99be23ef6ad tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1076.257793] env[63202]: DEBUG nova.network.neutron [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Successfully updated port: 76e223aa-1b1a-4d12-a8a6-89bb986656b5 {{(pid=63202) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1076.271563] env[63202]: DEBUG oslo_vmware.api [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52bdb80c-86aa-bd1d-8ee0-5879e607f1d5, 'name': SearchDatastore_Task, 'duration_secs': 0.012474} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.271856] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1076.272320] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore2] 9646810e-06d2-4de0-8f0c-e8719541da53/9646810e-06d2-4de0-8f0c-e8719541da53.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1076.272615] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4567c48a-d838-497f-ab79-57a14c076fff {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.280350] env[63202]: DEBUG oslo_vmware.api [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Waiting for the task: (returnval){ [ 1076.280350] env[63202]: value = "task-1385895" [ 1076.280350] env[63202]: _type = "Task" [ 1076.280350] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.288161] env[63202]: DEBUG oslo_vmware.api [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385895, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.654058] env[63202]: DEBUG nova.scheduler.client.report [None req-c3f9b308-d144-49dc-8dcc-b99be23ef6ad tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1076.761449] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquiring lock "refresh_cache-214c63b6-35e3-46e1-a9de-f3cea251c3c9" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1076.761788] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquired lock "refresh_cache-214c63b6-35e3-46e1-a9de-f3cea251c3c9" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1076.761988] env[63202]: DEBUG nova.network.neutron [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1076.790749] env[63202]: DEBUG oslo_vmware.api [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385895, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.976732] env[63202]: DEBUG nova.compute.manager [req-c3db9fa2-880c-4600-a465-f8b161c79bfa req-7e66fd02-fb34-4641-8128-cad332ca595a service nova] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Received event network-vif-plugged-76e223aa-1b1a-4d12-a8a6-89bb986656b5 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1076.978281] env[63202]: DEBUG oslo_concurrency.lockutils [req-c3db9fa2-880c-4600-a465-f8b161c79bfa req-7e66fd02-fb34-4641-8128-cad332ca595a service nova] Acquiring lock "214c63b6-35e3-46e1-a9de-f3cea251c3c9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1076.978281] env[63202]: DEBUG oslo_concurrency.lockutils [req-c3db9fa2-880c-4600-a465-f8b161c79bfa req-7e66fd02-fb34-4641-8128-cad332ca595a service nova] Lock "214c63b6-35e3-46e1-a9de-f3cea251c3c9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1076.978281] env[63202]: DEBUG oslo_concurrency.lockutils [req-c3db9fa2-880c-4600-a465-f8b161c79bfa req-7e66fd02-fb34-4641-8128-cad332ca595a service nova] Lock "214c63b6-35e3-46e1-a9de-f3cea251c3c9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1076.978281] env[63202]: DEBUG nova.compute.manager [req-c3db9fa2-880c-4600-a465-f8b161c79bfa req-7e66fd02-fb34-4641-8128-cad332ca595a service nova] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] No waiting events found dispatching network-vif-plugged-76e223aa-1b1a-4d12-a8a6-89bb986656b5 {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1076.978281] env[63202]: WARNING nova.compute.manager [req-c3db9fa2-880c-4600-a465-f8b161c79bfa req-7e66fd02-fb34-4641-8128-cad332ca595a service nova] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Received unexpected event network-vif-plugged-76e223aa-1b1a-4d12-a8a6-89bb986656b5 for instance with vm_state building and task_state spawning. [ 1076.978778] env[63202]: DEBUG nova.compute.manager [req-c3db9fa2-880c-4600-a465-f8b161c79bfa req-7e66fd02-fb34-4641-8128-cad332ca595a service nova] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Received event network-changed-76e223aa-1b1a-4d12-a8a6-89bb986656b5 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1076.978778] env[63202]: DEBUG nova.compute.manager [req-c3db9fa2-880c-4600-a465-f8b161c79bfa req-7e66fd02-fb34-4641-8128-cad332ca595a service nova] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Refreshing instance network info cache due to event network-changed-76e223aa-1b1a-4d12-a8a6-89bb986656b5. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1076.978778] env[63202]: DEBUG oslo_concurrency.lockutils [req-c3db9fa2-880c-4600-a465-f8b161c79bfa req-7e66fd02-fb34-4641-8128-cad332ca595a service nova] Acquiring lock "refresh_cache-214c63b6-35e3-46e1-a9de-f3cea251c3c9" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1077.158884] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c3f9b308-d144-49dc-8dcc-b99be23ef6ad tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.669s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.175377] env[63202]: INFO nova.scheduler.client.report [None req-c3f9b308-d144-49dc-8dcc-b99be23ef6ad tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Deleted allocations for instance 6e172a2a-1918-430c-84ce-005fef97bbf1 [ 1077.294353] env[63202]: DEBUG oslo_vmware.api [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385895, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.51385} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.294606] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore2] 9646810e-06d2-4de0-8f0c-e8719541da53/9646810e-06d2-4de0-8f0c-e8719541da53.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1077.294774] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1077.295069] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6fc7e682-50f4-4761-9f7e-0f2a1641066f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.302694] env[63202]: DEBUG oslo_vmware.api [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Waiting for the task: (returnval){ [ 1077.302694] env[63202]: value = "task-1385896" [ 1077.302694] env[63202]: _type = "Task" [ 1077.302694] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.303437] env[63202]: DEBUG nova.network.neutron [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1077.313167] env[63202]: DEBUG oslo_vmware.api [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385896, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.328371] env[63202]: DEBUG oslo_concurrency.lockutils [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "1cc050d2-1055-40b0-9769-dc03f7995da3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.328602] env[63202]: DEBUG oslo_concurrency.lockutils [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "1cc050d2-1055-40b0-9769-dc03f7995da3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.484500] env[63202]: DEBUG nova.network.neutron [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Updating instance_info_cache with network_info: [{"id": "76e223aa-1b1a-4d12-a8a6-89bb986656b5", "address": "fa:16:3e:8e:03:df", "network": {"id": "0aa55fee-953e-4c2d-b88a-43b0cf86c7f3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-878071414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7d28fea097541adadf0839940568409", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76e223aa-1b", "ovs_interfaceid": "76e223aa-1b1a-4d12-a8a6-89bb986656b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1077.682990] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c3f9b308-d144-49dc-8dcc-b99be23ef6ad tempest-ServerGroupTestJSON-1293124536 tempest-ServerGroupTestJSON-1293124536-project-member] Lock "6e172a2a-1918-430c-84ce-005fef97bbf1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.116s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.815359] env[63202]: DEBUG oslo_vmware.api [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385896, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082227} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.815623] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1077.816380] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73e0ae69-a321-45a0-84d7-dd4a387efa51 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.838366] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] 9646810e-06d2-4de0-8f0c-e8719541da53/9646810e-06d2-4de0-8f0c-e8719541da53.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1077.838716] env[63202]: DEBUG nova.compute.manager [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1077.841255] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-77eb7a5d-b00e-41ec-b73a-05eaaa955205 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.861134] env[63202]: DEBUG oslo_vmware.api [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Waiting for the task: (returnval){ [ 1077.861134] env[63202]: value = "task-1385897" [ 1077.861134] env[63202]: _type = "Task" [ 1077.861134] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.868913] env[63202]: DEBUG oslo_vmware.api [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385897, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.987536] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Releasing lock "refresh_cache-214c63b6-35e3-46e1-a9de-f3cea251c3c9" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1077.988088] env[63202]: DEBUG nova.compute.manager [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Instance network_info: |[{"id": "76e223aa-1b1a-4d12-a8a6-89bb986656b5", "address": "fa:16:3e:8e:03:df", "network": {"id": "0aa55fee-953e-4c2d-b88a-43b0cf86c7f3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-878071414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7d28fea097541adadf0839940568409", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76e223aa-1b", "ovs_interfaceid": "76e223aa-1b1a-4d12-a8a6-89bb986656b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1077.988498] env[63202]: DEBUG oslo_concurrency.lockutils [req-c3db9fa2-880c-4600-a465-f8b161c79bfa req-7e66fd02-fb34-4641-8128-cad332ca595a service nova] Acquired lock "refresh_cache-214c63b6-35e3-46e1-a9de-f3cea251c3c9" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.988714] env[63202]: DEBUG nova.network.neutron [req-c3db9fa2-880c-4600-a465-f8b161c79bfa req-7e66fd02-fb34-4641-8128-cad332ca595a service nova] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Refreshing network info cache for port 76e223aa-1b1a-4d12-a8a6-89bb986656b5 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1077.990051] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8e:03:df', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '76e223aa-1b1a-4d12-a8a6-89bb986656b5', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1077.998164] env[63202]: DEBUG oslo.service.loopingcall [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1077.999217] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1077.999449] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-caa7037d-5bcc-4271-9bc0-5d3bd789cf75 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.020057] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1078.020057] env[63202]: value = "task-1385898" [ 1078.020057] env[63202]: _type = "Task" [ 1078.020057] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.028469] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385898, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.372025] env[63202]: DEBUG oslo_vmware.api [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385897, 'name': ReconfigVM_Task, 'duration_secs': 0.29374} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.372287] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Reconfigured VM instance instance-00000066 to attach disk [datastore2] 9646810e-06d2-4de0-8f0c-e8719541da53/9646810e-06d2-4de0-8f0c-e8719541da53.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1078.372955] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d1ea91ba-99f3-439d-adbb-fd55954b0bb8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.376418] env[63202]: DEBUG oslo_concurrency.lockutils [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1078.376685] env[63202]: DEBUG oslo_concurrency.lockutils [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.378252] env[63202]: INFO nova.compute.claims [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1078.382192] env[63202]: DEBUG oslo_vmware.api [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Waiting for the task: (returnval){ [ 1078.382192] env[63202]: value = "task-1385899" [ 1078.382192] env[63202]: _type = "Task" [ 1078.382192] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.392137] env[63202]: DEBUG oslo_vmware.api [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385899, 'name': Rename_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.533952] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385898, 'name': CreateVM_Task, 'duration_secs': 0.302559} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.533952] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1078.533952] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1078.533952] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1078.533952] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1078.533952] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dcca4e6c-380a-48ad-8ff5-b4afc6fe2d62 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.539026] env[63202]: DEBUG oslo_vmware.api [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 1078.539026] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]521ac8a1-cefb-50e6-2994-a26a00f415cc" [ 1078.539026] env[63202]: _type = "Task" [ 1078.539026] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.545613] env[63202]: DEBUG oslo_vmware.api [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]521ac8a1-cefb-50e6-2994-a26a00f415cc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.818551] env[63202]: DEBUG nova.network.neutron [req-c3db9fa2-880c-4600-a465-f8b161c79bfa req-7e66fd02-fb34-4641-8128-cad332ca595a service nova] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Updated VIF entry in instance network info cache for port 76e223aa-1b1a-4d12-a8a6-89bb986656b5. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1078.818931] env[63202]: DEBUG nova.network.neutron [req-c3db9fa2-880c-4600-a465-f8b161c79bfa req-7e66fd02-fb34-4641-8128-cad332ca595a service nova] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Updating instance_info_cache with network_info: [{"id": "76e223aa-1b1a-4d12-a8a6-89bb986656b5", "address": "fa:16:3e:8e:03:df", "network": {"id": "0aa55fee-953e-4c2d-b88a-43b0cf86c7f3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-878071414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7d28fea097541adadf0839940568409", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76e223aa-1b", "ovs_interfaceid": "76e223aa-1b1a-4d12-a8a6-89bb986656b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1078.895499] env[63202]: DEBUG oslo_vmware.api [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385899, 'name': Rename_Task, 'duration_secs': 0.136954} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.895832] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1078.896126] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f62b3b1b-b52d-461f-b50e-a6cdf1d1b6f5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.903314] env[63202]: DEBUG oslo_vmware.api [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Waiting for the task: (returnval){ [ 1078.903314] env[63202]: value = "task-1385900" [ 1078.903314] env[63202]: _type = "Task" [ 1078.903314] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.912121] env[63202]: DEBUG oslo_vmware.api [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385900, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.048146] env[63202]: DEBUG oslo_vmware.api [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]521ac8a1-cefb-50e6-2994-a26a00f415cc, 'name': SearchDatastore_Task, 'duration_secs': 0.010481} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.048464] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1079.048692] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1079.048931] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1079.049093] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1079.049278] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1079.049554] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3c10361e-83a9-4976-8684-68955c6fb8a6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.058263] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1079.058263] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1079.058833] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-22dccf5a-dd1e-4821-a763-5ff6bc0e9d5b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.064199] env[63202]: DEBUG oslo_vmware.api [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 1079.064199] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52a5b246-6397-1c85-e7e1-c8cab4c9c823" [ 1079.064199] env[63202]: _type = "Task" [ 1079.064199] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.071782] env[63202]: DEBUG oslo_vmware.api [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52a5b246-6397-1c85-e7e1-c8cab4c9c823, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.322330] env[63202]: DEBUG oslo_concurrency.lockutils [req-c3db9fa2-880c-4600-a465-f8b161c79bfa req-7e66fd02-fb34-4641-8128-cad332ca595a service nova] Releasing lock "refresh_cache-214c63b6-35e3-46e1-a9de-f3cea251c3c9" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1079.419415] env[63202]: DEBUG oslo_vmware.api [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385900, 'name': PowerOnVM_Task, 'duration_secs': 0.50253} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.419856] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1079.419993] env[63202]: INFO nova.compute.manager [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Took 7.44 seconds to spawn the instance on the hypervisor. [ 1079.420488] env[63202]: DEBUG nova.compute.manager [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1079.421121] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-968623f1-9229-4cbe-8a07-1d531a4c8a5e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.483171] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97e887ac-3e38-4f23-a602-5cf28ed8d264 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.491288] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c080b09-6e6a-43a5-b34b-6b549b404994 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.523856] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1547b52-cdf2-41c1-acbe-e0f070aaf8f1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.531569] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa52aed9-36dd-4fd7-9a63-29882af6e592 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.545201] env[63202]: DEBUG nova.compute.provider_tree [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1079.575185] env[63202]: DEBUG oslo_vmware.api [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52a5b246-6397-1c85-e7e1-c8cab4c9c823, 'name': SearchDatastore_Task, 'duration_secs': 0.007887} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.575988] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0bcc9bc9-c022-4270-8f2a-558d9f89ce5b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.581458] env[63202]: DEBUG oslo_vmware.api [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 1079.581458] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52f6ed2d-86d5-4abb-5cef-afc7c05b7a64" [ 1079.581458] env[63202]: _type = "Task" [ 1079.581458] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.589144] env[63202]: DEBUG oslo_vmware.api [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52f6ed2d-86d5-4abb-5cef-afc7c05b7a64, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.939930] env[63202]: INFO nova.compute.manager [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Took 13.48 seconds to build instance. [ 1080.048945] env[63202]: DEBUG nova.scheduler.client.report [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1080.091990] env[63202]: DEBUG oslo_vmware.api [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52f6ed2d-86d5-4abb-5cef-afc7c05b7a64, 'name': SearchDatastore_Task, 'duration_secs': 0.008506} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.092279] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1080.092576] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore2] 214c63b6-35e3-46e1-a9de-f3cea251c3c9/214c63b6-35e3-46e1-a9de-f3cea251c3c9.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1080.092847] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f1f265fe-d694-416d-a792-fcd038341241 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.099355] env[63202]: DEBUG oslo_vmware.api [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 1080.099355] env[63202]: value = "task-1385901" [ 1080.099355] env[63202]: _type = "Task" [ 1080.099355] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.106870] env[63202]: DEBUG oslo_vmware.api [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385901, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.442553] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c1cca27e-6c3e-4e61-992d-8b93b323ea9d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Lock "9646810e-06d2-4de0-8f0c-e8719541da53" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.991s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1080.553826] env[63202]: DEBUG oslo_concurrency.lockutils [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.177s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1080.554443] env[63202]: DEBUG nova.compute.manager [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1080.609887] env[63202]: DEBUG oslo_vmware.api [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385901, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.059445] env[63202]: DEBUG nova.compute.utils [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1081.060867] env[63202]: DEBUG nova.compute.manager [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1081.061052] env[63202]: DEBUG nova.network.neutron [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1081.097933] env[63202]: DEBUG nova.policy [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b59e7e62d92e4f5eb0fee41e8f5196d7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f508ec04edc844a19640a8a85f27e5b9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 1081.109332] env[63202]: DEBUG oslo_vmware.api [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385901, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.649721} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.109599] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore2] 214c63b6-35e3-46e1-a9de-f3cea251c3c9/214c63b6-35e3-46e1-a9de-f3cea251c3c9.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1081.109843] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1081.110104] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f5150a50-55a5-46fc-8544-df4c5f90bc9e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.116676] env[63202]: DEBUG oslo_vmware.api [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 1081.116676] env[63202]: value = "task-1385902" [ 1081.116676] env[63202]: _type = "Task" [ 1081.116676] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.126064] env[63202]: DEBUG oslo_vmware.api [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385902, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.334199] env[63202]: DEBUG nova.network.neutron [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Successfully created port: ad8d36fb-afb1-480f-815c-6d60c7758b42 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1081.513305] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f5c81acc-bb51-46f9-9bb0-d850beb2f7a6 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Acquiring lock "e1e2e8a4-80aa-45eb-a90e-d20cba2943ee" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1081.513492] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f5c81acc-bb51-46f9-9bb0-d850beb2f7a6 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lock "e1e2e8a4-80aa-45eb-a90e-d20cba2943ee" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1081.564199] env[63202]: DEBUG nova.compute.manager [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1081.626048] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Acquiring lock "bed3be83-e467-4466-ad1f-9ca89bfa87a8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1081.626048] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Lock "bed3be83-e467-4466-ad1f-9ca89bfa87a8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1081.629914] env[63202]: DEBUG oslo_vmware.api [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385902, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062892} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.630179] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1081.630931] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6756a961-4c81-4068-807f-b53d5424ff3c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.653226] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] 214c63b6-35e3-46e1-a9de-f3cea251c3c9/214c63b6-35e3-46e1-a9de-f3cea251c3c9.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1081.653664] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-83c57cf2-6a34-4d33-b914-78b365f41569 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.673511] env[63202]: DEBUG oslo_vmware.api [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 1081.673511] env[63202]: value = "task-1385903" [ 1081.673511] env[63202]: _type = "Task" [ 1081.673511] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.681057] env[63202]: DEBUG oslo_vmware.api [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385903, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.017032] env[63202]: DEBUG nova.compute.utils [None req-f5c81acc-bb51-46f9-9bb0-d850beb2f7a6 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1082.128103] env[63202]: DEBUG nova.compute.manager [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1082.186603] env[63202]: DEBUG oslo_vmware.api [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385903, 'name': ReconfigVM_Task, 'duration_secs': 0.419596} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.187077] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Reconfigured VM instance instance-00000067 to attach disk [datastore2] 214c63b6-35e3-46e1-a9de-f3cea251c3c9/214c63b6-35e3-46e1-a9de-f3cea251c3c9.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1082.187888] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1baa21be-b097-4d9f-ba37-11b2ee793889 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.195961] env[63202]: DEBUG oslo_vmware.api [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 1082.195961] env[63202]: value = "task-1385904" [ 1082.195961] env[63202]: _type = "Task" [ 1082.195961] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.204413] env[63202]: DEBUG oslo_vmware.api [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385904, 'name': Rename_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.520525] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f5c81acc-bb51-46f9-9bb0-d850beb2f7a6 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lock "e1e2e8a4-80aa-45eb-a90e-d20cba2943ee" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.573797] env[63202]: DEBUG nova.compute.manager [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1082.598508] env[63202]: DEBUG nova.virt.hardware [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1082.598720] env[63202]: DEBUG nova.virt.hardware [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1082.598887] env[63202]: DEBUG nova.virt.hardware [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1082.599086] env[63202]: DEBUG nova.virt.hardware [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1082.599241] env[63202]: DEBUG nova.virt.hardware [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1082.599396] env[63202]: DEBUG nova.virt.hardware [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1082.599693] env[63202]: DEBUG nova.virt.hardware [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1082.599878] env[63202]: DEBUG nova.virt.hardware [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1082.600067] env[63202]: DEBUG nova.virt.hardware [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1082.600240] env[63202]: DEBUG nova.virt.hardware [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1082.600419] env[63202]: DEBUG nova.virt.hardware [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1082.601322] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af5a490b-25cd-47db-b761-293c4c196f86 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.608983] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f54443a2-8151-46c4-90dd-459ee7fe8c83 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.653306] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1082.653555] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.655141] env[63202]: INFO nova.compute.claims [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1082.705931] env[63202]: DEBUG oslo_vmware.api [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385904, 'name': Rename_Task, 'duration_secs': 0.166298} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.706175] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1082.706426] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1f9b70bb-c806-424d-a116-9854e678192b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.712931] env[63202]: DEBUG oslo_vmware.api [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 1082.712931] env[63202]: value = "task-1385905" [ 1082.712931] env[63202]: _type = "Task" [ 1082.712931] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.721218] env[63202]: DEBUG oslo_vmware.api [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385905, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.723187] env[63202]: DEBUG nova.compute.manager [req-108941d7-eaa8-440f-be50-efc8f7c3c61d req-19804ac6-4842-4c96-90a9-ad29fbf1cf73 service nova] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Received event network-vif-plugged-ad8d36fb-afb1-480f-815c-6d60c7758b42 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1082.723392] env[63202]: DEBUG oslo_concurrency.lockutils [req-108941d7-eaa8-440f-be50-efc8f7c3c61d req-19804ac6-4842-4c96-90a9-ad29fbf1cf73 service nova] Acquiring lock "1cc050d2-1055-40b0-9769-dc03f7995da3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1082.723599] env[63202]: DEBUG oslo_concurrency.lockutils [req-108941d7-eaa8-440f-be50-efc8f7c3c61d req-19804ac6-4842-4c96-90a9-ad29fbf1cf73 service nova] Lock "1cc050d2-1055-40b0-9769-dc03f7995da3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.723756] env[63202]: DEBUG oslo_concurrency.lockutils [req-108941d7-eaa8-440f-be50-efc8f7c3c61d req-19804ac6-4842-4c96-90a9-ad29fbf1cf73 service nova] Lock "1cc050d2-1055-40b0-9769-dc03f7995da3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.723926] env[63202]: DEBUG nova.compute.manager [req-108941d7-eaa8-440f-be50-efc8f7c3c61d req-19804ac6-4842-4c96-90a9-ad29fbf1cf73 service nova] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] No waiting events found dispatching network-vif-plugged-ad8d36fb-afb1-480f-815c-6d60c7758b42 {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1082.724145] env[63202]: WARNING nova.compute.manager [req-108941d7-eaa8-440f-be50-efc8f7c3c61d req-19804ac6-4842-4c96-90a9-ad29fbf1cf73 service nova] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Received unexpected event network-vif-plugged-ad8d36fb-afb1-480f-815c-6d60c7758b42 for instance with vm_state building and task_state spawning. [ 1082.809190] env[63202]: DEBUG nova.network.neutron [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Successfully updated port: ad8d36fb-afb1-480f-815c-6d60c7758b42 {{(pid=63202) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1083.223931] env[63202]: DEBUG oslo_vmware.api [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385905, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.313555] env[63202]: DEBUG oslo_concurrency.lockutils [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "refresh_cache-1cc050d2-1055-40b0-9769-dc03f7995da3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.313555] env[63202]: DEBUG oslo_concurrency.lockutils [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquired lock "refresh_cache-1cc050d2-1055-40b0-9769-dc03f7995da3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.313555] env[63202]: DEBUG nova.network.neutron [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1083.601076] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f5c81acc-bb51-46f9-9bb0-d850beb2f7a6 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Acquiring lock "e1e2e8a4-80aa-45eb-a90e-d20cba2943ee" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.601076] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f5c81acc-bb51-46f9-9bb0-d850beb2f7a6 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lock "e1e2e8a4-80aa-45eb-a90e-d20cba2943ee" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1083.601259] env[63202]: INFO nova.compute.manager [None req-f5c81acc-bb51-46f9-9bb0-d850beb2f7a6 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Attaching volume 9b6d6ca2-3b93-400f-aaa8-ed94ad761a49 to /dev/sdb [ 1083.631698] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07fdfe0b-a915-4445-961c-cf837b4f08ac {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.638736] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a1ce229-ec62-4e6a-ba8e-0860fedcca2a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.652467] env[63202]: DEBUG nova.virt.block_device [None req-f5c81acc-bb51-46f9-9bb0-d850beb2f7a6 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Updating existing volume attachment record: 214d532b-eb04-4fcf-97c8-1a96fc466d04 {{(pid=63202) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1083.722821] env[63202]: DEBUG oslo_vmware.api [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385905, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.743920] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-554fccfd-c27b-46de-b7ed-5e22c8216050 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.751141] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5182f9b-835d-4305-b6b5-320893f74c52 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.781203] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09433040-2af8-4128-81b7-faf87633fbf9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.789638] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a4da60f-de63-40eb-af60-348fb85132b8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.802226] env[63202]: DEBUG nova.compute.provider_tree [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1083.844729] env[63202]: DEBUG nova.network.neutron [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1084.018166] env[63202]: DEBUG nova.network.neutron [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Updating instance_info_cache with network_info: [{"id": "ad8d36fb-afb1-480f-815c-6d60c7758b42", "address": "fa:16:3e:47:0d:32", "network": {"id": "5a3b4c9b-2ca7-4f8b-8bbb-ea000db91402", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-543677750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f508ec04edc844a19640a8a85f27e5b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad8d36fb-af", "ovs_interfaceid": "ad8d36fb-afb1-480f-815c-6d60c7758b42", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1084.224083] env[63202]: DEBUG oslo_vmware.api [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385905, 'name': PowerOnVM_Task, 'duration_secs': 1.133169} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.224379] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1084.224566] env[63202]: INFO nova.compute.manager [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Took 8.39 seconds to spawn the instance on the hypervisor. [ 1084.224745] env[63202]: DEBUG nova.compute.manager [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1084.225546] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4a0da92-94db-4e7f-b51c-32415efa03bb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.304915] env[63202]: DEBUG nova.scheduler.client.report [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1084.521537] env[63202]: DEBUG oslo_concurrency.lockutils [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Releasing lock "refresh_cache-1cc050d2-1055-40b0-9769-dc03f7995da3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1084.521887] env[63202]: DEBUG nova.compute.manager [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Instance network_info: |[{"id": "ad8d36fb-afb1-480f-815c-6d60c7758b42", "address": "fa:16:3e:47:0d:32", "network": {"id": "5a3b4c9b-2ca7-4f8b-8bbb-ea000db91402", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-543677750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f508ec04edc844a19640a8a85f27e5b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad8d36fb-af", "ovs_interfaceid": "ad8d36fb-afb1-480f-815c-6d60c7758b42", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1084.522351] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:47:0d:32', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ad8d36fb-afb1-480f-815c-6d60c7758b42', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1084.529936] env[63202]: DEBUG oslo.service.loopingcall [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1084.530167] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1084.530399] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-00629412-3add-4779-abe7-33981f96d458 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.551038] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1084.551038] env[63202]: value = "task-1385907" [ 1084.551038] env[63202]: _type = "Task" [ 1084.551038] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.558503] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385907, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.742611] env[63202]: INFO nova.compute.manager [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Took 13.33 seconds to build instance. [ 1084.748742] env[63202]: DEBUG nova.compute.manager [req-9aed5cde-dc74-4dca-9760-a9ff458ecb84 req-6d42b592-0833-40ce-953f-0a1252e74573 service nova] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Received event network-changed-ad8d36fb-afb1-480f-815c-6d60c7758b42 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1084.748985] env[63202]: DEBUG nova.compute.manager [req-9aed5cde-dc74-4dca-9760-a9ff458ecb84 req-6d42b592-0833-40ce-953f-0a1252e74573 service nova] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Refreshing instance network info cache due to event network-changed-ad8d36fb-afb1-480f-815c-6d60c7758b42. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1084.749476] env[63202]: DEBUG oslo_concurrency.lockutils [req-9aed5cde-dc74-4dca-9760-a9ff458ecb84 req-6d42b592-0833-40ce-953f-0a1252e74573 service nova] Acquiring lock "refresh_cache-1cc050d2-1055-40b0-9769-dc03f7995da3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1084.749476] env[63202]: DEBUG oslo_concurrency.lockutils [req-9aed5cde-dc74-4dca-9760-a9ff458ecb84 req-6d42b592-0833-40ce-953f-0a1252e74573 service nova] Acquired lock "refresh_cache-1cc050d2-1055-40b0-9769-dc03f7995da3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1084.749843] env[63202]: DEBUG nova.network.neutron [req-9aed5cde-dc74-4dca-9760-a9ff458ecb84 req-6d42b592-0833-40ce-953f-0a1252e74573 service nova] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Refreshing network info cache for port ad8d36fb-afb1-480f-815c-6d60c7758b42 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1084.810588] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.157s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.811358] env[63202]: DEBUG nova.compute.manager [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1085.061107] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385907, 'name': CreateVM_Task, 'duration_secs': 0.288791} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.061306] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1085.062044] env[63202]: DEBUG oslo_concurrency.lockutils [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1085.062241] env[63202]: DEBUG oslo_concurrency.lockutils [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1085.062599] env[63202]: DEBUG oslo_concurrency.lockutils [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1085.062893] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce6c7f97-b939-47e2-b220-635b02d0d620 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.067520] env[63202]: DEBUG oslo_vmware.api [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 1085.067520] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]521bd3ef-2f28-4c95-c9b7-dc7e4723850e" [ 1085.067520] env[63202]: _type = "Task" [ 1085.067520] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.075020] env[63202]: DEBUG oslo_vmware.api [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]521bd3ef-2f28-4c95-c9b7-dc7e4723850e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.244471] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dc22684f-b9bc-4990-a72a-efca9aebff5d tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "214c63b6-35e3-46e1-a9de-f3cea251c3c9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.839s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1085.317024] env[63202]: DEBUG nova.compute.utils [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1085.318476] env[63202]: DEBUG nova.compute.manager [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1085.318476] env[63202]: DEBUG nova.network.neutron [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1085.359652] env[63202]: DEBUG nova.policy [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '47a40b09812d4be68bb478771514dfa3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '433b1d11832147b886bed7a3a4952768', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 1085.543070] env[63202]: DEBUG nova.network.neutron [req-9aed5cde-dc74-4dca-9760-a9ff458ecb84 req-6d42b592-0833-40ce-953f-0a1252e74573 service nova] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Updated VIF entry in instance network info cache for port ad8d36fb-afb1-480f-815c-6d60c7758b42. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1085.543440] env[63202]: DEBUG nova.network.neutron [req-9aed5cde-dc74-4dca-9760-a9ff458ecb84 req-6d42b592-0833-40ce-953f-0a1252e74573 service nova] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Updating instance_info_cache with network_info: [{"id": "ad8d36fb-afb1-480f-815c-6d60c7758b42", "address": "fa:16:3e:47:0d:32", "network": {"id": "5a3b4c9b-2ca7-4f8b-8bbb-ea000db91402", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-543677750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f508ec04edc844a19640a8a85f27e5b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad8d36fb-af", "ovs_interfaceid": "ad8d36fb-afb1-480f-815c-6d60c7758b42", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1085.589141] env[63202]: DEBUG oslo_vmware.api [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]521bd3ef-2f28-4c95-c9b7-dc7e4723850e, 'name': SearchDatastore_Task, 'duration_secs': 0.009481} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.589141] env[63202]: DEBUG oslo_concurrency.lockutils [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1085.589141] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1085.589141] env[63202]: DEBUG oslo_concurrency.lockutils [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1085.589384] env[63202]: DEBUG oslo_concurrency.lockutils [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1085.589384] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1085.589384] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4d5a8a0f-b9bb-40e3-8e64-5741b483274d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.601372] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1085.601372] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1085.601372] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-402d2a0e-5d62-4eeb-9bc0-2c97d71accd7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.610127] env[63202]: DEBUG oslo_vmware.api [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 1085.610127] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52ffb720-962a-458d-af0d-dff567a9d08f" [ 1085.610127] env[63202]: _type = "Task" [ 1085.610127] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.625945] env[63202]: DEBUG oslo_vmware.api [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52ffb720-962a-458d-af0d-dff567a9d08f, 'name': SearchDatastore_Task, 'duration_secs': 0.008942} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.626769] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1593863a-a679-418c-a7ad-09abb1759fcd {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.631795] env[63202]: DEBUG oslo_vmware.api [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 1085.631795] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52a3d587-02cd-af9e-4e55-b87e85c4b70b" [ 1085.631795] env[63202]: _type = "Task" [ 1085.631795] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.638916] env[63202]: DEBUG oslo_vmware.api [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52a3d587-02cd-af9e-4e55-b87e85c4b70b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.821430] env[63202]: DEBUG nova.compute.manager [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1085.949418] env[63202]: DEBUG nova.network.neutron [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Successfully created port: 17f1d8d2-ca08-4427-9bf3-73551a56829d {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1086.046221] env[63202]: DEBUG oslo_concurrency.lockutils [req-9aed5cde-dc74-4dca-9760-a9ff458ecb84 req-6d42b592-0833-40ce-953f-0a1252e74573 service nova] Releasing lock "refresh_cache-1cc050d2-1055-40b0-9769-dc03f7995da3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1086.142671] env[63202]: DEBUG oslo_vmware.api [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52a3d587-02cd-af9e-4e55-b87e85c4b70b, 'name': SearchDatastore_Task, 'duration_secs': 0.00788} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.142994] env[63202]: DEBUG oslo_concurrency.lockutils [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1086.143313] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore2] 1cc050d2-1055-40b0-9769-dc03f7995da3/1cc050d2-1055-40b0-9769-dc03f7995da3.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1086.143592] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-684ca5d5-55eb-4cb6-ae30-c2052c2d8b27 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.150259] env[63202]: DEBUG oslo_vmware.api [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 1086.150259] env[63202]: value = "task-1385909" [ 1086.150259] env[63202]: _type = "Task" [ 1086.150259] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.157263] env[63202]: DEBUG oslo_vmware.api [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385909, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.398521] env[63202]: DEBUG nova.compute.manager [req-64a95257-637a-4e78-8a82-7647b29f4a54 req-a8621ae6-657c-4ecf-8030-8dd1ddafc869 service nova] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Received event network-changed-76e223aa-1b1a-4d12-a8a6-89bb986656b5 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1086.398794] env[63202]: DEBUG nova.compute.manager [req-64a95257-637a-4e78-8a82-7647b29f4a54 req-a8621ae6-657c-4ecf-8030-8dd1ddafc869 service nova] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Refreshing instance network info cache due to event network-changed-76e223aa-1b1a-4d12-a8a6-89bb986656b5. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1086.399020] env[63202]: DEBUG oslo_concurrency.lockutils [req-64a95257-637a-4e78-8a82-7647b29f4a54 req-a8621ae6-657c-4ecf-8030-8dd1ddafc869 service nova] Acquiring lock "refresh_cache-214c63b6-35e3-46e1-a9de-f3cea251c3c9" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1086.399250] env[63202]: DEBUG oslo_concurrency.lockutils [req-64a95257-637a-4e78-8a82-7647b29f4a54 req-a8621ae6-657c-4ecf-8030-8dd1ddafc869 service nova] Acquired lock "refresh_cache-214c63b6-35e3-46e1-a9de-f3cea251c3c9" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.399435] env[63202]: DEBUG nova.network.neutron [req-64a95257-637a-4e78-8a82-7647b29f4a54 req-a8621ae6-657c-4ecf-8030-8dd1ddafc869 service nova] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Refreshing network info cache for port 76e223aa-1b1a-4d12-a8a6-89bb986656b5 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1086.663412] env[63202]: DEBUG oslo_vmware.api [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385909, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.477433} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.663412] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore2] 1cc050d2-1055-40b0-9769-dc03f7995da3/1cc050d2-1055-40b0-9769-dc03f7995da3.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1086.663412] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1086.663412] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c96b12a0-8d68-422b-aba4-90b4b5d3fd75 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.669020] env[63202]: DEBUG oslo_vmware.api [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 1086.669020] env[63202]: value = "task-1385910" [ 1086.669020] env[63202]: _type = "Task" [ 1086.669020] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.676153] env[63202]: DEBUG oslo_vmware.api [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385910, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.833124] env[63202]: DEBUG nova.compute.manager [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1086.859034] env[63202]: DEBUG nova.virt.hardware [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1086.859375] env[63202]: DEBUG nova.virt.hardware [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1086.859582] env[63202]: DEBUG nova.virt.hardware [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1086.859824] env[63202]: DEBUG nova.virt.hardware [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1086.860032] env[63202]: DEBUG nova.virt.hardware [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1086.860387] env[63202]: DEBUG nova.virt.hardware [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1086.860467] env[63202]: DEBUG nova.virt.hardware [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1086.860709] env[63202]: DEBUG nova.virt.hardware [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1086.860897] env[63202]: DEBUG nova.virt.hardware [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1086.861086] env[63202]: DEBUG nova.virt.hardware [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1086.861261] env[63202]: DEBUG nova.virt.hardware [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1086.862159] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6408d3c-e6f9-46d7-837b-97e86a293db6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.869874] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f538b65-7bcb-41f6-98e7-3da4ceca2add {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.110143] env[63202]: DEBUG nova.network.neutron [req-64a95257-637a-4e78-8a82-7647b29f4a54 req-a8621ae6-657c-4ecf-8030-8dd1ddafc869 service nova] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Updated VIF entry in instance network info cache for port 76e223aa-1b1a-4d12-a8a6-89bb986656b5. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1087.110515] env[63202]: DEBUG nova.network.neutron [req-64a95257-637a-4e78-8a82-7647b29f4a54 req-a8621ae6-657c-4ecf-8030-8dd1ddafc869 service nova] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Updating instance_info_cache with network_info: [{"id": "76e223aa-1b1a-4d12-a8a6-89bb986656b5", "address": "fa:16:3e:8e:03:df", "network": {"id": "0aa55fee-953e-4c2d-b88a-43b0cf86c7f3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-878071414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.211", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7d28fea097541adadf0839940568409", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76e223aa-1b", "ovs_interfaceid": "76e223aa-1b1a-4d12-a8a6-89bb986656b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1087.177073] env[63202]: DEBUG oslo_vmware.api [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385910, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064317} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.177332] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1087.178132] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdabec1b-a677-44a2-9958-f436169b8491 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.200278] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] 1cc050d2-1055-40b0-9769-dc03f7995da3/1cc050d2-1055-40b0-9769-dc03f7995da3.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1087.200567] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3c4b2ba3-4fe2-4a30-a419-31b064131881 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.220839] env[63202]: DEBUG oslo_vmware.api [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 1087.220839] env[63202]: value = "task-1385911" [ 1087.220839] env[63202]: _type = "Task" [ 1087.220839] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.230387] env[63202]: DEBUG oslo_vmware.api [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385911, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.613556] env[63202]: DEBUG oslo_concurrency.lockutils [req-64a95257-637a-4e78-8a82-7647b29f4a54 req-a8621ae6-657c-4ecf-8030-8dd1ddafc869 service nova] Releasing lock "refresh_cache-214c63b6-35e3-46e1-a9de-f3cea251c3c9" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1087.622147] env[63202]: DEBUG nova.network.neutron [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Successfully updated port: 17f1d8d2-ca08-4427-9bf3-73551a56829d {{(pid=63202) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1087.731497] env[63202]: DEBUG oslo_vmware.api [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385911, 'name': ReconfigVM_Task, 'duration_secs': 0.392463} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.733032] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Reconfigured VM instance instance-00000068 to attach disk [datastore2] 1cc050d2-1055-40b0-9769-dc03f7995da3/1cc050d2-1055-40b0-9769-dc03f7995da3.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1087.733032] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-beca89ee-f968-4af5-b546-72a2ba15afed {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.738753] env[63202]: DEBUG oslo_vmware.api [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 1087.738753] env[63202]: value = "task-1385912" [ 1087.738753] env[63202]: _type = "Task" [ 1087.738753] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.746782] env[63202]: DEBUG oslo_vmware.api [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385912, 'name': Rename_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.124763] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Acquiring lock "refresh_cache-bed3be83-e467-4466-ad1f-9ca89bfa87a8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1088.124979] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Acquired lock "refresh_cache-bed3be83-e467-4466-ad1f-9ca89bfa87a8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1088.125200] env[63202]: DEBUG nova.network.neutron [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1088.196321] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-f5c81acc-bb51-46f9-9bb0-d850beb2f7a6 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Volume attach. Driver type: vmdk {{(pid=63202) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1088.196566] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-f5c81acc-bb51-46f9-9bb0-d850beb2f7a6 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294225', 'volume_id': '9b6d6ca2-3b93-400f-aaa8-ed94ad761a49', 'name': 'volume-9b6d6ca2-3b93-400f-aaa8-ed94ad761a49', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e1e2e8a4-80aa-45eb-a90e-d20cba2943ee', 'attached_at': '', 'detached_at': '', 'volume_id': '9b6d6ca2-3b93-400f-aaa8-ed94ad761a49', 'serial': '9b6d6ca2-3b93-400f-aaa8-ed94ad761a49'} {{(pid=63202) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1088.197449] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b1c6435-6a9b-4047-a169-a84904dabecf {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.215281] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2da5cb1-d690-4c01-bb34-c1c9fcd8a88d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.238985] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-f5c81acc-bb51-46f9-9bb0-d850beb2f7a6 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] volume-9b6d6ca2-3b93-400f-aaa8-ed94ad761a49/volume-9b6d6ca2-3b93-400f-aaa8-ed94ad761a49.vmdk or device None with type thin {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1088.239273] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a9dff7d0-f5c1-4ba4-9a82-9a11288db0aa {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.259367] env[63202]: DEBUG oslo_vmware.api [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385912, 'name': Rename_Task, 'duration_secs': 0.132929} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.260829] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1088.261166] env[63202]: DEBUG oslo_vmware.api [None req-f5c81acc-bb51-46f9-9bb0-d850beb2f7a6 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Waiting for the task: (returnval){ [ 1088.261166] env[63202]: value = "task-1385913" [ 1088.261166] env[63202]: _type = "Task" [ 1088.261166] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.261347] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3ffff8c7-2627-4096-b7e5-7914dbeab1ff {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.270211] env[63202]: DEBUG oslo_vmware.api [None req-f5c81acc-bb51-46f9-9bb0-d850beb2f7a6 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385913, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.271308] env[63202]: DEBUG oslo_vmware.api [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 1088.271308] env[63202]: value = "task-1385914" [ 1088.271308] env[63202]: _type = "Task" [ 1088.271308] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.278182] env[63202]: DEBUG oslo_vmware.api [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385914, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.426799] env[63202]: DEBUG nova.compute.manager [req-9784738f-e241-4584-834b-f271d6f30ac2 req-ec88650f-d6fe-442c-a65f-7ed0020bf2cf service nova] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Received event network-vif-plugged-17f1d8d2-ca08-4427-9bf3-73551a56829d {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1088.427113] env[63202]: DEBUG oslo_concurrency.lockutils [req-9784738f-e241-4584-834b-f271d6f30ac2 req-ec88650f-d6fe-442c-a65f-7ed0020bf2cf service nova] Acquiring lock "bed3be83-e467-4466-ad1f-9ca89bfa87a8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1088.427357] env[63202]: DEBUG oslo_concurrency.lockutils [req-9784738f-e241-4584-834b-f271d6f30ac2 req-ec88650f-d6fe-442c-a65f-7ed0020bf2cf service nova] Lock "bed3be83-e467-4466-ad1f-9ca89bfa87a8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1088.427561] env[63202]: DEBUG oslo_concurrency.lockutils [req-9784738f-e241-4584-834b-f271d6f30ac2 req-ec88650f-d6fe-442c-a65f-7ed0020bf2cf service nova] Lock "bed3be83-e467-4466-ad1f-9ca89bfa87a8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.427776] env[63202]: DEBUG nova.compute.manager [req-9784738f-e241-4584-834b-f271d6f30ac2 req-ec88650f-d6fe-442c-a65f-7ed0020bf2cf service nova] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] No waiting events found dispatching network-vif-plugged-17f1d8d2-ca08-4427-9bf3-73551a56829d {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1088.428017] env[63202]: WARNING nova.compute.manager [req-9784738f-e241-4584-834b-f271d6f30ac2 req-ec88650f-d6fe-442c-a65f-7ed0020bf2cf service nova] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Received unexpected event network-vif-plugged-17f1d8d2-ca08-4427-9bf3-73551a56829d for instance with vm_state building and task_state spawning. [ 1088.428273] env[63202]: DEBUG nova.compute.manager [req-9784738f-e241-4584-834b-f271d6f30ac2 req-ec88650f-d6fe-442c-a65f-7ed0020bf2cf service nova] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Received event network-changed-17f1d8d2-ca08-4427-9bf3-73551a56829d {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1088.428469] env[63202]: DEBUG nova.compute.manager [req-9784738f-e241-4584-834b-f271d6f30ac2 req-ec88650f-d6fe-442c-a65f-7ed0020bf2cf service nova] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Refreshing instance network info cache due to event network-changed-17f1d8d2-ca08-4427-9bf3-73551a56829d. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1088.428668] env[63202]: DEBUG oslo_concurrency.lockutils [req-9784738f-e241-4584-834b-f271d6f30ac2 req-ec88650f-d6fe-442c-a65f-7ed0020bf2cf service nova] Acquiring lock "refresh_cache-bed3be83-e467-4466-ad1f-9ca89bfa87a8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1088.668965] env[63202]: DEBUG nova.network.neutron [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1088.773172] env[63202]: DEBUG oslo_vmware.api [None req-f5c81acc-bb51-46f9-9bb0-d850beb2f7a6 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385913, 'name': ReconfigVM_Task, 'duration_secs': 0.461857} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.776260] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-f5c81acc-bb51-46f9-9bb0-d850beb2f7a6 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Reconfigured VM instance instance-00000062 to attach disk [datastore2] volume-9b6d6ca2-3b93-400f-aaa8-ed94ad761a49/volume-9b6d6ca2-3b93-400f-aaa8-ed94ad761a49.vmdk or device None with type thin {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1088.781295] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b2a48212-cff4-40b9-8a55-81609dae73cd {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.802217] env[63202]: DEBUG oslo_vmware.api [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385914, 'name': PowerOnVM_Task, 'duration_secs': 0.472591} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.803481] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1088.803695] env[63202]: INFO nova.compute.manager [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Took 6.23 seconds to spawn the instance on the hypervisor. [ 1088.803876] env[63202]: DEBUG nova.compute.manager [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1088.804226] env[63202]: DEBUG oslo_vmware.api [None req-f5c81acc-bb51-46f9-9bb0-d850beb2f7a6 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Waiting for the task: (returnval){ [ 1088.804226] env[63202]: value = "task-1385915" [ 1088.804226] env[63202]: _type = "Task" [ 1088.804226] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.804919] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d0f0c08-14cd-4fd6-99b0-0ff2e938de13 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.819905] env[63202]: DEBUG oslo_vmware.api [None req-f5c81acc-bb51-46f9-9bb0-d850beb2f7a6 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385915, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.827715] env[63202]: DEBUG nova.network.neutron [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Updating instance_info_cache with network_info: [{"id": "17f1d8d2-ca08-4427-9bf3-73551a56829d", "address": "fa:16:3e:c3:3a:21", "network": {"id": "f105b723-b2a9-401d-a936-e31b4b6609f6", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-298435423-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "433b1d11832147b886bed7a3a4952768", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e4c8c8fd-baca-4e60-97dc-ff0418d63215", "external-id": "nsx-vlan-transportzone-178", "segmentation_id": 178, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap17f1d8d2-ca", "ovs_interfaceid": "17f1d8d2-ca08-4427-9bf3-73551a56829d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1089.317975] env[63202]: DEBUG oslo_vmware.api [None req-f5c81acc-bb51-46f9-9bb0-d850beb2f7a6 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385915, 'name': ReconfigVM_Task, 'duration_secs': 0.140963} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.318333] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-f5c81acc-bb51-46f9-9bb0-d850beb2f7a6 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294225', 'volume_id': '9b6d6ca2-3b93-400f-aaa8-ed94ad761a49', 'name': 'volume-9b6d6ca2-3b93-400f-aaa8-ed94ad761a49', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e1e2e8a4-80aa-45eb-a90e-d20cba2943ee', 'attached_at': '', 'detached_at': '', 'volume_id': '9b6d6ca2-3b93-400f-aaa8-ed94ad761a49', 'serial': '9b6d6ca2-3b93-400f-aaa8-ed94ad761a49'} {{(pid=63202) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1089.327482] env[63202]: INFO nova.compute.manager [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Took 10.97 seconds to build instance. [ 1089.329853] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Releasing lock "refresh_cache-bed3be83-e467-4466-ad1f-9ca89bfa87a8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1089.330118] env[63202]: DEBUG nova.compute.manager [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Instance network_info: |[{"id": "17f1d8d2-ca08-4427-9bf3-73551a56829d", "address": "fa:16:3e:c3:3a:21", "network": {"id": "f105b723-b2a9-401d-a936-e31b4b6609f6", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-298435423-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "433b1d11832147b886bed7a3a4952768", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e4c8c8fd-baca-4e60-97dc-ff0418d63215", "external-id": "nsx-vlan-transportzone-178", "segmentation_id": 178, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap17f1d8d2-ca", "ovs_interfaceid": "17f1d8d2-ca08-4427-9bf3-73551a56829d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1089.330399] env[63202]: DEBUG oslo_concurrency.lockutils [req-9784738f-e241-4584-834b-f271d6f30ac2 req-ec88650f-d6fe-442c-a65f-7ed0020bf2cf service nova] Acquired lock "refresh_cache-bed3be83-e467-4466-ad1f-9ca89bfa87a8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1089.330573] env[63202]: DEBUG nova.network.neutron [req-9784738f-e241-4584-834b-f271d6f30ac2 req-ec88650f-d6fe-442c-a65f-7ed0020bf2cf service nova] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Refreshing network info cache for port 17f1d8d2-ca08-4427-9bf3-73551a56829d {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1089.331515] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c3:3a:21', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e4c8c8fd-baca-4e60-97dc-ff0418d63215', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '17f1d8d2-ca08-4427-9bf3-73551a56829d', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1089.339672] env[63202]: DEBUG oslo.service.loopingcall [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1089.341489] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1089.341774] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-081c46d4-4155-443b-9735-f2afa11db44b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.363694] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1089.363694] env[63202]: value = "task-1385916" [ 1089.363694] env[63202]: _type = "Task" [ 1089.363694] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.371958] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385916, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.829669] env[63202]: DEBUG oslo_concurrency.lockutils [None req-693a281c-9da3-428c-a182-117f002b0e35 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "1cc050d2-1055-40b0-9769-dc03f7995da3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.501s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.873763] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385916, 'name': CreateVM_Task} progress is 99%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.031635] env[63202]: DEBUG nova.network.neutron [req-9784738f-e241-4584-834b-f271d6f30ac2 req-ec88650f-d6fe-442c-a65f-7ed0020bf2cf service nova] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Updated VIF entry in instance network info cache for port 17f1d8d2-ca08-4427-9bf3-73551a56829d. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1090.032132] env[63202]: DEBUG nova.network.neutron [req-9784738f-e241-4584-834b-f271d6f30ac2 req-ec88650f-d6fe-442c-a65f-7ed0020bf2cf service nova] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Updating instance_info_cache with network_info: [{"id": "17f1d8d2-ca08-4427-9bf3-73551a56829d", "address": "fa:16:3e:c3:3a:21", "network": {"id": "f105b723-b2a9-401d-a936-e31b4b6609f6", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-298435423-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "433b1d11832147b886bed7a3a4952768", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e4c8c8fd-baca-4e60-97dc-ff0418d63215", "external-id": "nsx-vlan-transportzone-178", "segmentation_id": 178, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap17f1d8d2-ca", "ovs_interfaceid": "17f1d8d2-ca08-4427-9bf3-73551a56829d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1090.353233] env[63202]: DEBUG nova.objects.instance [None req-f5c81acc-bb51-46f9-9bb0-d850beb2f7a6 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lazy-loading 'flavor' on Instance uuid e1e2e8a4-80aa-45eb-a90e-d20cba2943ee {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1090.374843] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385916, 'name': CreateVM_Task} progress is 99%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.535703] env[63202]: DEBUG oslo_concurrency.lockutils [req-9784738f-e241-4584-834b-f271d6f30ac2 req-ec88650f-d6fe-442c-a65f-7ed0020bf2cf service nova] Releasing lock "refresh_cache-bed3be83-e467-4466-ad1f-9ca89bfa87a8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1090.859081] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f5c81acc-bb51-46f9-9bb0-d850beb2f7a6 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lock "e1e2e8a4-80aa-45eb-a90e-d20cba2943ee" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.257s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.874291] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385916, 'name': CreateVM_Task, 'duration_secs': 1.368771} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.874456] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1090.875121] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1090.875292] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1090.875613] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1090.875869] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3671c48f-6459-4af8-88c8-3629ac240ef6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.880400] env[63202]: DEBUG oslo_vmware.api [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Waiting for the task: (returnval){ [ 1090.880400] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52f78f4b-16cb-e341-82a7-f0d733a934e6" [ 1090.880400] env[63202]: _type = "Task" [ 1090.880400] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.887851] env[63202]: DEBUG oslo_vmware.api [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52f78f4b-16cb-e341-82a7-f0d733a934e6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.226084] env[63202]: DEBUG nova.compute.manager [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Stashing vm_state: active {{(pid=63202) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1091.391040] env[63202]: DEBUG oslo_vmware.api [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52f78f4b-16cb-e341-82a7-f0d733a934e6, 'name': SearchDatastore_Task, 'duration_secs': 0.01068} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.391310] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1091.391548] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1091.391827] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1091.391993] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1091.392209] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1091.392484] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-04eea066-685f-4591-a38d-a4bac47a9091 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.401545] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1091.401545] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1091.401962] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d07d5280-829c-43a4-8e29-6524abca4bc7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.407248] env[63202]: DEBUG oslo_vmware.api [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Waiting for the task: (returnval){ [ 1091.407248] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52d97b4e-f473-3899-b941-220da3efdfb1" [ 1091.407248] env[63202]: _type = "Task" [ 1091.407248] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.415451] env[63202]: DEBUG oslo_vmware.api [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52d97b4e-f473-3899-b941-220da3efdfb1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.746698] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.746976] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.819341] env[63202]: DEBUG oslo_concurrency.lockutils [None req-805b6188-f2f2-44a7-987b-cb08f79877bc tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Acquiring lock "e1e2e8a4-80aa-45eb-a90e-d20cba2943ee" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.819571] env[63202]: DEBUG oslo_concurrency.lockutils [None req-805b6188-f2f2-44a7-987b-cb08f79877bc tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lock "e1e2e8a4-80aa-45eb-a90e-d20cba2943ee" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.919308] env[63202]: DEBUG oslo_vmware.api [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52d97b4e-f473-3899-b941-220da3efdfb1, 'name': SearchDatastore_Task, 'duration_secs': 0.009713} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.920092] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf82d6cb-05e0-493d-997e-95e96cb8d5f5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.925303] env[63202]: DEBUG oslo_vmware.api [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Waiting for the task: (returnval){ [ 1091.925303] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52d5bb99-e5e6-c56b-77c1-e074b30a07c1" [ 1091.925303] env[63202]: _type = "Task" [ 1091.925303] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.933151] env[63202]: DEBUG oslo_vmware.api [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52d5bb99-e5e6-c56b-77c1-e074b30a07c1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.252372] env[63202]: INFO nova.compute.claims [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1092.323927] env[63202]: DEBUG nova.compute.utils [None req-805b6188-f2f2-44a7-987b-cb08f79877bc tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1092.435944] env[63202]: DEBUG oslo_vmware.api [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52d5bb99-e5e6-c56b-77c1-e074b30a07c1, 'name': SearchDatastore_Task, 'duration_secs': 0.012116} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.436231] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1092.436491] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore2] bed3be83-e467-4466-ad1f-9ca89bfa87a8/bed3be83-e467-4466-ad1f-9ca89bfa87a8.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1092.436738] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bc94d452-03cd-4595-aadb-b8be77cfcac9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.443837] env[63202]: DEBUG oslo_vmware.api [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Waiting for the task: (returnval){ [ 1092.443837] env[63202]: value = "task-1385917" [ 1092.443837] env[63202]: _type = "Task" [ 1092.443837] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.451681] env[63202]: DEBUG oslo_vmware.api [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385917, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.759289] env[63202]: INFO nova.compute.resource_tracker [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Updating resource usage from migration d0fa269f-50a0-48dc-b06a-d3d25ba42c25 [ 1092.826888] env[63202]: DEBUG oslo_concurrency.lockutils [None req-805b6188-f2f2-44a7-987b-cb08f79877bc tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lock "e1e2e8a4-80aa-45eb-a90e-d20cba2943ee" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1092.863772] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-776117a0-4afd-4bbd-857d-b7bae6c1d8ae {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.873407] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc82a9d1-5dd9-4390-8b4a-15763c8c64f7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.907462] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac091a14-6784-434b-bac4-83a96a869fd2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.916035] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5fd0696-3bcb-4b0a-ad5d-dc86ccec46bb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.930407] env[63202]: DEBUG nova.compute.provider_tree [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Updating inventory in ProviderTree for provider 79b33d17-3e75-494c-a550-67b275de2079 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1092.952989] env[63202]: DEBUG oslo_vmware.api [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385917, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.450026] env[63202]: ERROR nova.scheduler.client.report [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [req-9c7e7fcc-a2a4-4260-b2e4-2cfcdf69e8fb] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 79b33d17-3e75-494c-a550-67b275de2079. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-9c7e7fcc-a2a4-4260-b2e4-2cfcdf69e8fb"}]} [ 1093.457535] env[63202]: DEBUG oslo_vmware.api [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385917, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.638665} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.457958] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore2] bed3be83-e467-4466-ad1f-9ca89bfa87a8/bed3be83-e467-4466-ad1f-9ca89bfa87a8.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1093.458360] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1093.459024] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2fc0e2c3-ba75-4a2e-aa2f-8bf0acb07a6a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.465891] env[63202]: DEBUG oslo_vmware.api [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Waiting for the task: (returnval){ [ 1093.465891] env[63202]: value = "task-1385918" [ 1093.465891] env[63202]: _type = "Task" [ 1093.465891] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.469450] env[63202]: DEBUG nova.scheduler.client.report [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Refreshing inventories for resource provider 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1093.476910] env[63202]: DEBUG oslo_vmware.api [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385918, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.483975] env[63202]: DEBUG nova.scheduler.client.report [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Updating ProviderTree inventory for provider 79b33d17-3e75-494c-a550-67b275de2079 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1093.484292] env[63202]: DEBUG nova.compute.provider_tree [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Updating inventory in ProviderTree for provider 79b33d17-3e75-494c-a550-67b275de2079 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1093.567168] env[63202]: DEBUG nova.scheduler.client.report [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Refreshing aggregate associations for resource provider 79b33d17-3e75-494c-a550-67b275de2079, aggregates: None {{(pid=63202) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1093.589019] env[63202]: DEBUG nova.scheduler.client.report [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Refreshing trait associations for resource provider 79b33d17-3e75-494c-a550-67b275de2079, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,HW_ARCH_X86_64,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO {{(pid=63202) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1093.670956] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c24b674-8c4a-42ef-b768-0f22ec3cc0f5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.678528] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f819f0fd-4d27-4f9c-a7e7-3aa6a483defa {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.708248] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0862403d-6bcc-442b-8635-6922eec6f72e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.715269] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0020f69-6101-4934-ac58-580a9262fff4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.727728] env[63202]: DEBUG nova.compute.provider_tree [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Updating inventory in ProviderTree for provider 79b33d17-3e75-494c-a550-67b275de2079 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1093.888013] env[63202]: DEBUG oslo_concurrency.lockutils [None req-805b6188-f2f2-44a7-987b-cb08f79877bc tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Acquiring lock "e1e2e8a4-80aa-45eb-a90e-d20cba2943ee" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.888235] env[63202]: DEBUG oslo_concurrency.lockutils [None req-805b6188-f2f2-44a7-987b-cb08f79877bc tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lock "e1e2e8a4-80aa-45eb-a90e-d20cba2943ee" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1093.888444] env[63202]: INFO nova.compute.manager [None req-805b6188-f2f2-44a7-987b-cb08f79877bc tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Attaching volume d200389d-4a94-4c63-8aa4-4d2b298bf88f to /dev/sdc [ 1093.918207] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3471ee41-b1a4-4a6d-b3da-d809faf89ffd {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.924991] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15633c46-f843-484b-aacf-2a2e4fd30839 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.937853] env[63202]: DEBUG nova.virt.block_device [None req-805b6188-f2f2-44a7-987b-cb08f79877bc tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Updating existing volume attachment record: 4d108fe8-5ac4-44fb-afcc-01de237e9328 {{(pid=63202) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1093.975387] env[63202]: DEBUG oslo_vmware.api [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385918, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06624} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.975638] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1093.976394] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7631030-ce1c-46eb-9c28-8cfa3ff28a85 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.997940] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] bed3be83-e467-4466-ad1f-9ca89bfa87a8/bed3be83-e467-4466-ad1f-9ca89bfa87a8.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1093.998517] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c068b22a-259e-44e5-9877-0774e197038f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.018819] env[63202]: DEBUG oslo_vmware.api [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Waiting for the task: (returnval){ [ 1094.018819] env[63202]: value = "task-1385919" [ 1094.018819] env[63202]: _type = "Task" [ 1094.018819] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.026172] env[63202]: DEBUG oslo_vmware.api [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385919, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.256769] env[63202]: DEBUG nova.scheduler.client.report [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Updated inventory for provider 79b33d17-3e75-494c-a550-67b275de2079 with generation 130 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1094.257646] env[63202]: DEBUG nova.compute.provider_tree [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Updating resource provider 79b33d17-3e75-494c-a550-67b275de2079 generation from 130 to 131 during operation: update_inventory {{(pid=63202) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1094.257646] env[63202]: DEBUG nova.compute.provider_tree [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Updating inventory in ProviderTree for provider 79b33d17-3e75-494c-a550-67b275de2079 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1094.528857] env[63202]: DEBUG oslo_vmware.api [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385919, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.762789] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 3.016s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1094.762981] env[63202]: INFO nova.compute.manager [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Migrating [ 1095.030796] env[63202]: DEBUG oslo_vmware.api [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385919, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.277684] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "refresh_cache-1cc050d2-1055-40b0-9769-dc03f7995da3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1095.277965] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquired lock "refresh_cache-1cc050d2-1055-40b0-9769-dc03f7995da3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1095.278295] env[63202]: DEBUG nova.network.neutron [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1095.529229] env[63202]: DEBUG oslo_vmware.api [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385919, 'name': ReconfigVM_Task, 'duration_secs': 1.25299} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.529513] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Reconfigured VM instance instance-00000069 to attach disk [datastore2] bed3be83-e467-4466-ad1f-9ca89bfa87a8/bed3be83-e467-4466-ad1f-9ca89bfa87a8.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1095.530168] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3e151ebf-e2aa-4b09-af98-516399f853d5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.535762] env[63202]: DEBUG oslo_vmware.api [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Waiting for the task: (returnval){ [ 1095.535762] env[63202]: value = "task-1385921" [ 1095.535762] env[63202]: _type = "Task" [ 1095.535762] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.542921] env[63202]: DEBUG oslo_vmware.api [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385921, 'name': Rename_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.969604] env[63202]: DEBUG nova.network.neutron [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Updating instance_info_cache with network_info: [{"id": "ad8d36fb-afb1-480f-815c-6d60c7758b42", "address": "fa:16:3e:47:0d:32", "network": {"id": "5a3b4c9b-2ca7-4f8b-8bbb-ea000db91402", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-543677750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f508ec04edc844a19640a8a85f27e5b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad8d36fb-af", "ovs_interfaceid": "ad8d36fb-afb1-480f-815c-6d60c7758b42", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1096.046427] env[63202]: DEBUG oslo_vmware.api [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385921, 'name': Rename_Task, 'duration_secs': 0.134831} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.046762] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1096.046949] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e4a2067b-71a0-4b17-bdb7-d334397da89c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.052874] env[63202]: DEBUG oslo_vmware.api [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Waiting for the task: (returnval){ [ 1096.052874] env[63202]: value = "task-1385922" [ 1096.052874] env[63202]: _type = "Task" [ 1096.052874] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.060163] env[63202]: DEBUG oslo_vmware.api [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385922, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.472761] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Releasing lock "refresh_cache-1cc050d2-1055-40b0-9769-dc03f7995da3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1096.562896] env[63202]: DEBUG oslo_vmware.api [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385922, 'name': PowerOnVM_Task, 'duration_secs': 0.462026} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.563177] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1096.563381] env[63202]: INFO nova.compute.manager [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Took 9.73 seconds to spawn the instance on the hypervisor. [ 1096.563561] env[63202]: DEBUG nova.compute.manager [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1096.564408] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-403952a8-6e1c-486e-af93-140e7a7c40ac {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.079262] env[63202]: INFO nova.compute.manager [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Took 14.45 seconds to build instance. [ 1097.582272] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a41f7a06-5d5b-4a2b-aa5b-c3634fd4f580 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Lock "bed3be83-e467-4466-ad1f-9ca89bfa87a8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.957s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1097.987327] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-532afbbe-774f-49d9-8c05-ec81aaa0c218 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.007028] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Updating instance '1cc050d2-1055-40b0-9769-dc03f7995da3' progress to 0 {{(pid=63202) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1098.162812] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a1a4dd99-aa1a-48fc-ad14-7387f5a1eb6b tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Acquiring lock "bed3be83-e467-4466-ad1f-9ca89bfa87a8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1098.162812] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a1a4dd99-aa1a-48fc-ad14-7387f5a1eb6b tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Lock "bed3be83-e467-4466-ad1f-9ca89bfa87a8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1098.162812] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a1a4dd99-aa1a-48fc-ad14-7387f5a1eb6b tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Acquiring lock "bed3be83-e467-4466-ad1f-9ca89bfa87a8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1098.163263] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a1a4dd99-aa1a-48fc-ad14-7387f5a1eb6b tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Lock "bed3be83-e467-4466-ad1f-9ca89bfa87a8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1098.163263] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a1a4dd99-aa1a-48fc-ad14-7387f5a1eb6b tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Lock "bed3be83-e467-4466-ad1f-9ca89bfa87a8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1098.165303] env[63202]: INFO nova.compute.manager [None req-a1a4dd99-aa1a-48fc-ad14-7387f5a1eb6b tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Terminating instance [ 1098.167021] env[63202]: DEBUG nova.compute.manager [None req-a1a4dd99-aa1a-48fc-ad14-7387f5a1eb6b tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1098.167221] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a1a4dd99-aa1a-48fc-ad14-7387f5a1eb6b tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1098.168070] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c35155c-ceea-4b36-96f6-4694b65532f3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.176159] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1a4dd99-aa1a-48fc-ad14-7387f5a1eb6b tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1098.176404] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4f63b3fe-349a-4a1a-a986-10f8dd6bb042 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.182635] env[63202]: DEBUG oslo_vmware.api [None req-a1a4dd99-aa1a-48fc-ad14-7387f5a1eb6b tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Waiting for the task: (returnval){ [ 1098.182635] env[63202]: value = "task-1385924" [ 1098.182635] env[63202]: _type = "Task" [ 1098.182635] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.190614] env[63202]: DEBUG oslo_vmware.api [None req-a1a4dd99-aa1a-48fc-ad14-7387f5a1eb6b tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385924, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.480991] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-805b6188-f2f2-44a7-987b-cb08f79877bc tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Volume attach. Driver type: vmdk {{(pid=63202) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1098.481261] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-805b6188-f2f2-44a7-987b-cb08f79877bc tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294228', 'volume_id': 'd200389d-4a94-4c63-8aa4-4d2b298bf88f', 'name': 'volume-d200389d-4a94-4c63-8aa4-4d2b298bf88f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e1e2e8a4-80aa-45eb-a90e-d20cba2943ee', 'attached_at': '', 'detached_at': '', 'volume_id': 'd200389d-4a94-4c63-8aa4-4d2b298bf88f', 'serial': 'd200389d-4a94-4c63-8aa4-4d2b298bf88f'} {{(pid=63202) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1098.482215] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12f3f7d1-9963-4d89-9903-21a5d366762e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.498372] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65a57336-823d-476a-b292-100401745428 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.521360] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1098.528965] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-805b6188-f2f2-44a7-987b-cb08f79877bc tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] volume-d200389d-4a94-4c63-8aa4-4d2b298bf88f/volume-d200389d-4a94-4c63-8aa4-4d2b298bf88f.vmdk or device None with type thin {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1098.529254] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ebda69b1-3879-40a3-83e9-97a61cc89335 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.530842] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ff7477d3-c37c-4222-ad20-d7af704985fc {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.549722] env[63202]: DEBUG oslo_vmware.api [None req-805b6188-f2f2-44a7-987b-cb08f79877bc tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Waiting for the task: (returnval){ [ 1098.549722] env[63202]: value = "task-1385926" [ 1098.549722] env[63202]: _type = "Task" [ 1098.549722] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.551123] env[63202]: DEBUG oslo_vmware.api [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 1098.551123] env[63202]: value = "task-1385925" [ 1098.551123] env[63202]: _type = "Task" [ 1098.551123] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.566498] env[63202]: DEBUG oslo_vmware.api [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385925, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.566742] env[63202]: DEBUG oslo_vmware.api [None req-805b6188-f2f2-44a7-987b-cb08f79877bc tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385926, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.692866] env[63202]: DEBUG oslo_vmware.api [None req-a1a4dd99-aa1a-48fc-ad14-7387f5a1eb6b tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385924, 'name': PowerOffVM_Task, 'duration_secs': 0.17145} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.693289] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1a4dd99-aa1a-48fc-ad14-7387f5a1eb6b tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1098.693511] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a1a4dd99-aa1a-48fc-ad14-7387f5a1eb6b tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1098.693845] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f5fb11fa-9216-40cb-ac09-79f0696b0d93 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.783255] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a1a4dd99-aa1a-48fc-ad14-7387f5a1eb6b tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1098.783482] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a1a4dd99-aa1a-48fc-ad14-7387f5a1eb6b tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Deleting contents of the VM from datastore datastore2 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1098.783681] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1a4dd99-aa1a-48fc-ad14-7387f5a1eb6b tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Deleting the datastore file [datastore2] bed3be83-e467-4466-ad1f-9ca89bfa87a8 {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1098.783943] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-440337a2-1997-438e-9ced-43bdf34e1ad2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.794170] env[63202]: DEBUG oslo_vmware.api [None req-a1a4dd99-aa1a-48fc-ad14-7387f5a1eb6b tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Waiting for the task: (returnval){ [ 1098.794170] env[63202]: value = "task-1385928" [ 1098.794170] env[63202]: _type = "Task" [ 1098.794170] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.801811] env[63202]: DEBUG oslo_vmware.api [None req-a1a4dd99-aa1a-48fc-ad14-7387f5a1eb6b tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385928, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.064611] env[63202]: DEBUG oslo_vmware.api [None req-805b6188-f2f2-44a7-987b-cb08f79877bc tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385926, 'name': ReconfigVM_Task, 'duration_secs': 0.361416} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.067489] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-805b6188-f2f2-44a7-987b-cb08f79877bc tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Reconfigured VM instance instance-00000062 to attach disk [datastore1] volume-d200389d-4a94-4c63-8aa4-4d2b298bf88f/volume-d200389d-4a94-4c63-8aa4-4d2b298bf88f.vmdk or device None with type thin {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1099.072105] env[63202]: DEBUG oslo_vmware.api [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385925, 'name': PowerOffVM_Task, 'duration_secs': 0.226132} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.072324] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ed6c970a-1556-4626-a68f-42ced7d9cd03 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.081699] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1099.081917] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Updating instance '1cc050d2-1055-40b0-9769-dc03f7995da3' progress to 17 {{(pid=63202) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1099.090613] env[63202]: DEBUG oslo_vmware.api [None req-805b6188-f2f2-44a7-987b-cb08f79877bc tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Waiting for the task: (returnval){ [ 1099.090613] env[63202]: value = "task-1385929" [ 1099.090613] env[63202]: _type = "Task" [ 1099.090613] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.098206] env[63202]: DEBUG oslo_vmware.api [None req-805b6188-f2f2-44a7-987b-cb08f79877bc tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385929, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.305300] env[63202]: DEBUG oslo_vmware.api [None req-a1a4dd99-aa1a-48fc-ad14-7387f5a1eb6b tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385928, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139908} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.305558] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1a4dd99-aa1a-48fc-ad14-7387f5a1eb6b tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1099.305741] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a1a4dd99-aa1a-48fc-ad14-7387f5a1eb6b tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Deleted contents of the VM from datastore datastore2 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1099.306037] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a1a4dd99-aa1a-48fc-ad14-7387f5a1eb6b tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1099.306108] env[63202]: INFO nova.compute.manager [None req-a1a4dd99-aa1a-48fc-ad14-7387f5a1eb6b tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1099.306338] env[63202]: DEBUG oslo.service.loopingcall [None req-a1a4dd99-aa1a-48fc-ad14-7387f5a1eb6b tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1099.306590] env[63202]: DEBUG nova.compute.manager [-] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1099.306694] env[63202]: DEBUG nova.network.neutron [-] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1099.558205] env[63202]: DEBUG nova.compute.manager [req-307aab6a-fdb6-4d85-8354-fb707b897d3a req-0783e628-bc1d-49cd-9831-94a76adc2032 service nova] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Received event network-vif-deleted-17f1d8d2-ca08-4427-9bf3-73551a56829d {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1099.558405] env[63202]: INFO nova.compute.manager [req-307aab6a-fdb6-4d85-8354-fb707b897d3a req-0783e628-bc1d-49cd-9831-94a76adc2032 service nova] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Neutron deleted interface 17f1d8d2-ca08-4427-9bf3-73551a56829d; detaching it from the instance and deleting it from the info cache [ 1099.558577] env[63202]: DEBUG nova.network.neutron [req-307aab6a-fdb6-4d85-8354-fb707b897d3a req-0783e628-bc1d-49cd-9831-94a76adc2032 service nova] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1099.588536] env[63202]: DEBUG nova.virt.hardware [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1099.588806] env[63202]: DEBUG nova.virt.hardware [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1099.589230] env[63202]: DEBUG nova.virt.hardware [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1099.589230] env[63202]: DEBUG nova.virt.hardware [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1099.589363] env[63202]: DEBUG nova.virt.hardware [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1099.589568] env[63202]: DEBUG nova.virt.hardware [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1099.589934] env[63202]: DEBUG nova.virt.hardware [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1099.590057] env[63202]: DEBUG nova.virt.hardware [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1099.590247] env[63202]: DEBUG nova.virt.hardware [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1099.590417] env[63202]: DEBUG nova.virt.hardware [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1099.590592] env[63202]: DEBUG nova.virt.hardware [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1099.596150] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7a0c51b0-1e5f-4476-8eaa-32b10ff96f2a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.614660] env[63202]: DEBUG oslo_vmware.api [None req-805b6188-f2f2-44a7-987b-cb08f79877bc tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385929, 'name': ReconfigVM_Task, 'duration_secs': 0.139993} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.618252] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-805b6188-f2f2-44a7-987b-cb08f79877bc tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294228', 'volume_id': 'd200389d-4a94-4c63-8aa4-4d2b298bf88f', 'name': 'volume-d200389d-4a94-4c63-8aa4-4d2b298bf88f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e1e2e8a4-80aa-45eb-a90e-d20cba2943ee', 'attached_at': '', 'detached_at': '', 'volume_id': 'd200389d-4a94-4c63-8aa4-4d2b298bf88f', 'serial': 'd200389d-4a94-4c63-8aa4-4d2b298bf88f'} {{(pid=63202) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1099.618557] env[63202]: DEBUG oslo_vmware.api [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 1099.618557] env[63202]: value = "task-1385930" [ 1099.618557] env[63202]: _type = "Task" [ 1099.618557] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.626631] env[63202]: DEBUG oslo_vmware.api [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385930, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.032474] env[63202]: DEBUG nova.network.neutron [-] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1100.060924] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8fa13e49-f808-4532-bce7-45b69c20ab04 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.070488] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40a1b387-0127-4b34-8f0b-44ce3a870690 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.097989] env[63202]: DEBUG nova.compute.manager [req-307aab6a-fdb6-4d85-8354-fb707b897d3a req-0783e628-bc1d-49cd-9831-94a76adc2032 service nova] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Detach interface failed, port_id=17f1d8d2-ca08-4427-9bf3-73551a56829d, reason: Instance bed3be83-e467-4466-ad1f-9ca89bfa87a8 could not be found. {{(pid=63202) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1100.129760] env[63202]: DEBUG oslo_vmware.api [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385930, 'name': ReconfigVM_Task, 'duration_secs': 0.191637} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.130088] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Updating instance '1cc050d2-1055-40b0-9769-dc03f7995da3' progress to 33 {{(pid=63202) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1100.536031] env[63202]: INFO nova.compute.manager [-] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Took 1.23 seconds to deallocate network for instance. [ 1100.637229] env[63202]: DEBUG nova.virt.hardware [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1100.637509] env[63202]: DEBUG nova.virt.hardware [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1100.637670] env[63202]: DEBUG nova.virt.hardware [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1100.637855] env[63202]: DEBUG nova.virt.hardware [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1100.638009] env[63202]: DEBUG nova.virt.hardware [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1100.638167] env[63202]: DEBUG nova.virt.hardware [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1100.638372] env[63202]: DEBUG nova.virt.hardware [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1100.638531] env[63202]: DEBUG nova.virt.hardware [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1100.638697] env[63202]: DEBUG nova.virt.hardware [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1100.638860] env[63202]: DEBUG nova.virt.hardware [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1100.639044] env[63202]: DEBUG nova.virt.hardware [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1100.644447] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Reconfiguring VM instance instance-00000068 to detach disk 2000 {{(pid=63202) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1100.644728] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cb6a2f40-4d22-4d37-9d31-b55f77bac43a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.658450] env[63202]: DEBUG nova.objects.instance [None req-805b6188-f2f2-44a7-987b-cb08f79877bc tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lazy-loading 'flavor' on Instance uuid e1e2e8a4-80aa-45eb-a90e-d20cba2943ee {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1100.665814] env[63202]: DEBUG oslo_vmware.api [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 1100.665814] env[63202]: value = "task-1385931" [ 1100.665814] env[63202]: _type = "Task" [ 1100.665814] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.673550] env[63202]: DEBUG oslo_vmware.api [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385931, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.042221] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a1a4dd99-aa1a-48fc-ad14-7387f5a1eb6b tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1101.042517] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a1a4dd99-aa1a-48fc-ad14-7387f5a1eb6b tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1101.042745] env[63202]: DEBUG nova.objects.instance [None req-a1a4dd99-aa1a-48fc-ad14-7387f5a1eb6b tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Lazy-loading 'resources' on Instance uuid bed3be83-e467-4466-ad1f-9ca89bfa87a8 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1101.163760] env[63202]: DEBUG oslo_concurrency.lockutils [None req-805b6188-f2f2-44a7-987b-cb08f79877bc tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lock "e1e2e8a4-80aa-45eb-a90e-d20cba2943ee" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.275s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1101.175464] env[63202]: DEBUG oslo_vmware.api [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385931, 'name': ReconfigVM_Task, 'duration_secs': 0.151084} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.176237] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Reconfigured VM instance instance-00000068 to detach disk 2000 {{(pid=63202) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1101.177012] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eca601b0-ea1e-42db-9fcc-d7455d0afeac {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.199015] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] 1cc050d2-1055-40b0-9769-dc03f7995da3/1cc050d2-1055-40b0-9769-dc03f7995da3.vmdk or device None with type thin {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1101.199479] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4766b9b8-5d56-46a8-94cf-a1b6914db393 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.217395] env[63202]: DEBUG oslo_vmware.api [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 1101.217395] env[63202]: value = "task-1385932" [ 1101.217395] env[63202]: _type = "Task" [ 1101.217395] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.225424] env[63202]: DEBUG oslo_vmware.api [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385932, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.453631] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dfe099dc-ccbd-4a88-a2c5-494dd564ae34 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Acquiring lock "e1e2e8a4-80aa-45eb-a90e-d20cba2943ee" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1101.453900] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dfe099dc-ccbd-4a88-a2c5-494dd564ae34 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lock "e1e2e8a4-80aa-45eb-a90e-d20cba2943ee" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1101.635364] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c59adf2-b8f9-4ce9-b8f7-8d5c869645b5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.642930] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b314b34d-f331-4a2d-afee-16b1b379c0ce {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.673183] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d43f189-c3b9-43c0-91c5-5ffefd82acbe {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.680161] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a497420-d8be-4619-bb9b-5771744eb9de {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.692889] env[63202]: DEBUG nova.compute.provider_tree [None req-a1a4dd99-aa1a-48fc-ad14-7387f5a1eb6b tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Updating inventory in ProviderTree for provider 79b33d17-3e75-494c-a550-67b275de2079 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1101.726660] env[63202]: DEBUG oslo_vmware.api [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385932, 'name': ReconfigVM_Task, 'duration_secs': 0.256162} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.726875] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Reconfigured VM instance instance-00000068 to attach disk [datastore2] 1cc050d2-1055-40b0-9769-dc03f7995da3/1cc050d2-1055-40b0-9769-dc03f7995da3.vmdk or device None with type thin {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1101.727213] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Updating instance '1cc050d2-1055-40b0-9769-dc03f7995da3' progress to 50 {{(pid=63202) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1101.957600] env[63202]: INFO nova.compute.manager [None req-dfe099dc-ccbd-4a88-a2c5-494dd564ae34 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Detaching volume 9b6d6ca2-3b93-400f-aaa8-ed94ad761a49 [ 1101.987850] env[63202]: INFO nova.virt.block_device [None req-dfe099dc-ccbd-4a88-a2c5-494dd564ae34 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Attempting to driver detach volume 9b6d6ca2-3b93-400f-aaa8-ed94ad761a49 from mountpoint /dev/sdb [ 1101.988110] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-dfe099dc-ccbd-4a88-a2c5-494dd564ae34 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Volume detach. Driver type: vmdk {{(pid=63202) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1101.988320] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-dfe099dc-ccbd-4a88-a2c5-494dd564ae34 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294225', 'volume_id': '9b6d6ca2-3b93-400f-aaa8-ed94ad761a49', 'name': 'volume-9b6d6ca2-3b93-400f-aaa8-ed94ad761a49', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e1e2e8a4-80aa-45eb-a90e-d20cba2943ee', 'attached_at': '', 'detached_at': '', 'volume_id': '9b6d6ca2-3b93-400f-aaa8-ed94ad761a49', 'serial': '9b6d6ca2-3b93-400f-aaa8-ed94ad761a49'} {{(pid=63202) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1101.989219] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-387c7263-2614-4172-9ad2-fd415e7aed6d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.012560] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4907587b-8519-4b2e-b5d9-5a05f57c1331 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.019739] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de6ea4a2-a9f1-4d88-8580-7d9dede7799e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.042955] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a2f2fb5-c69a-4298-8fd4-e2c1295446ef {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.057206] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-dfe099dc-ccbd-4a88-a2c5-494dd564ae34 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] The volume has not been displaced from its original location: [datastore2] volume-9b6d6ca2-3b93-400f-aaa8-ed94ad761a49/volume-9b6d6ca2-3b93-400f-aaa8-ed94ad761a49.vmdk. No consolidation needed. {{(pid=63202) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1102.062427] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-dfe099dc-ccbd-4a88-a2c5-494dd564ae34 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Reconfiguring VM instance instance-00000062 to detach disk 2001 {{(pid=63202) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1102.062689] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7609b1dc-d789-4446-8cda-bc2a81dde3f4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.080673] env[63202]: DEBUG oslo_vmware.api [None req-dfe099dc-ccbd-4a88-a2c5-494dd564ae34 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Waiting for the task: (returnval){ [ 1102.080673] env[63202]: value = "task-1385933" [ 1102.080673] env[63202]: _type = "Task" [ 1102.080673] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.087982] env[63202]: DEBUG oslo_vmware.api [None req-dfe099dc-ccbd-4a88-a2c5-494dd564ae34 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385933, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.224528] env[63202]: DEBUG nova.scheduler.client.report [None req-a1a4dd99-aa1a-48fc-ad14-7387f5a1eb6b tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Updated inventory for provider 79b33d17-3e75-494c-a550-67b275de2079 with generation 131 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1102.224815] env[63202]: DEBUG nova.compute.provider_tree [None req-a1a4dd99-aa1a-48fc-ad14-7387f5a1eb6b tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Updating resource provider 79b33d17-3e75-494c-a550-67b275de2079 generation from 131 to 132 during operation: update_inventory {{(pid=63202) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1102.225016] env[63202]: DEBUG nova.compute.provider_tree [None req-a1a4dd99-aa1a-48fc-ad14-7387f5a1eb6b tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Updating inventory in ProviderTree for provider 79b33d17-3e75-494c-a550-67b275de2079 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1102.233049] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f603f4a-68ca-4910-b8ea-fa03289606af {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.252351] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edab8230-9f1b-4363-b12b-14ac3d9609f4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.270628] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Updating instance '1cc050d2-1055-40b0-9769-dc03f7995da3' progress to 67 {{(pid=63202) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1102.590305] env[63202]: DEBUG oslo_vmware.api [None req-dfe099dc-ccbd-4a88-a2c5-494dd564ae34 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385933, 'name': ReconfigVM_Task, 'duration_secs': 0.248511} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.590523] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-dfe099dc-ccbd-4a88-a2c5-494dd564ae34 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Reconfigured VM instance instance-00000062 to detach disk 2001 {{(pid=63202) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1102.595219] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-811e344a-7135-45df-bd7f-e053d9d79d76 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.610068] env[63202]: DEBUG oslo_vmware.api [None req-dfe099dc-ccbd-4a88-a2c5-494dd564ae34 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Waiting for the task: (returnval){ [ 1102.610068] env[63202]: value = "task-1385934" [ 1102.610068] env[63202]: _type = "Task" [ 1102.610068] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.619246] env[63202]: DEBUG oslo_vmware.api [None req-dfe099dc-ccbd-4a88-a2c5-494dd564ae34 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385934, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.730451] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a1a4dd99-aa1a-48fc-ad14-7387f5a1eb6b tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.688s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.748233] env[63202]: INFO nova.scheduler.client.report [None req-a1a4dd99-aa1a-48fc-ad14-7387f5a1eb6b tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Deleted allocations for instance bed3be83-e467-4466-ad1f-9ca89bfa87a8 [ 1102.810698] env[63202]: DEBUG nova.network.neutron [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Port ad8d36fb-afb1-480f-815c-6d60c7758b42 binding to destination host cpu-1 is already ACTIVE {{(pid=63202) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1103.120288] env[63202]: DEBUG oslo_vmware.api [None req-dfe099dc-ccbd-4a88-a2c5-494dd564ae34 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385934, 'name': ReconfigVM_Task, 'duration_secs': 0.127605} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.120658] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-dfe099dc-ccbd-4a88-a2c5-494dd564ae34 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294225', 'volume_id': '9b6d6ca2-3b93-400f-aaa8-ed94ad761a49', 'name': 'volume-9b6d6ca2-3b93-400f-aaa8-ed94ad761a49', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e1e2e8a4-80aa-45eb-a90e-d20cba2943ee', 'attached_at': '', 'detached_at': '', 'volume_id': '9b6d6ca2-3b93-400f-aaa8-ed94ad761a49', 'serial': '9b6d6ca2-3b93-400f-aaa8-ed94ad761a49'} {{(pid=63202) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1103.256013] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a1a4dd99-aa1a-48fc-ad14-7387f5a1eb6b tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Lock "bed3be83-e467-4466-ad1f-9ca89bfa87a8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.093s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1103.661436] env[63202]: DEBUG nova.objects.instance [None req-dfe099dc-ccbd-4a88-a2c5-494dd564ae34 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lazy-loading 'flavor' on Instance uuid e1e2e8a4-80aa-45eb-a90e-d20cba2943ee {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1103.831209] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "1cc050d2-1055-40b0-9769-dc03f7995da3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1103.831479] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "1cc050d2-1055-40b0-9769-dc03f7995da3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1103.831609] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "1cc050d2-1055-40b0-9769-dc03f7995da3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1104.668445] env[63202]: DEBUG oslo_concurrency.lockutils [None req-dfe099dc-ccbd-4a88-a2c5-494dd564ae34 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lock "e1e2e8a4-80aa-45eb-a90e-d20cba2943ee" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.214s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1104.690478] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2026e6a5-90f0-487b-95f0-c2b50f0a7e24 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Acquiring lock "e1e2e8a4-80aa-45eb-a90e-d20cba2943ee" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1104.690725] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2026e6a5-90f0-487b-95f0-c2b50f0a7e24 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lock "e1e2e8a4-80aa-45eb-a90e-d20cba2943ee" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1104.867212] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "refresh_cache-1cc050d2-1055-40b0-9769-dc03f7995da3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1104.867428] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquired lock "refresh_cache-1cc050d2-1055-40b0-9769-dc03f7995da3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1104.867621] env[63202]: DEBUG nova.network.neutron [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1105.193359] env[63202]: INFO nova.compute.manager [None req-2026e6a5-90f0-487b-95f0-c2b50f0a7e24 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Detaching volume d200389d-4a94-4c63-8aa4-4d2b298bf88f [ 1105.223103] env[63202]: INFO nova.virt.block_device [None req-2026e6a5-90f0-487b-95f0-c2b50f0a7e24 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Attempting to driver detach volume d200389d-4a94-4c63-8aa4-4d2b298bf88f from mountpoint /dev/sdc [ 1105.223364] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-2026e6a5-90f0-487b-95f0-c2b50f0a7e24 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Volume detach. Driver type: vmdk {{(pid=63202) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1105.223556] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-2026e6a5-90f0-487b-95f0-c2b50f0a7e24 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294228', 'volume_id': 'd200389d-4a94-4c63-8aa4-4d2b298bf88f', 'name': 'volume-d200389d-4a94-4c63-8aa4-4d2b298bf88f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e1e2e8a4-80aa-45eb-a90e-d20cba2943ee', 'attached_at': '', 'detached_at': '', 'volume_id': 'd200389d-4a94-4c63-8aa4-4d2b298bf88f', 'serial': 'd200389d-4a94-4c63-8aa4-4d2b298bf88f'} {{(pid=63202) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1105.224471] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc050422-3df1-48d4-b635-f586c10362bc {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.246885] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2f0347a-3c91-489f-9523-01af439019be {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.253529] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08288e12-21a3-47e4-b83d-e35fa7a593f3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.272738] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9937decc-2142-440d-972e-f7dcde933018 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.286654] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-2026e6a5-90f0-487b-95f0-c2b50f0a7e24 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] The volume has not been displaced from its original location: [datastore1] volume-d200389d-4a94-4c63-8aa4-4d2b298bf88f/volume-d200389d-4a94-4c63-8aa4-4d2b298bf88f.vmdk. No consolidation needed. {{(pid=63202) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1105.291877] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-2026e6a5-90f0-487b-95f0-c2b50f0a7e24 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Reconfiguring VM instance instance-00000062 to detach disk 2002 {{(pid=63202) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1105.292201] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d196a931-145e-4d19-bf57-a959b8653509 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.309138] env[63202]: DEBUG oslo_vmware.api [None req-2026e6a5-90f0-487b-95f0-c2b50f0a7e24 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Waiting for the task: (returnval){ [ 1105.309138] env[63202]: value = "task-1385935" [ 1105.309138] env[63202]: _type = "Task" [ 1105.309138] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.571592] env[63202]: DEBUG nova.network.neutron [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Updating instance_info_cache with network_info: [{"id": "ad8d36fb-afb1-480f-815c-6d60c7758b42", "address": "fa:16:3e:47:0d:32", "network": {"id": "5a3b4c9b-2ca7-4f8b-8bbb-ea000db91402", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-543677750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f508ec04edc844a19640a8a85f27e5b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad8d36fb-af", "ovs_interfaceid": "ad8d36fb-afb1-480f-815c-6d60c7758b42", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1105.818988] env[63202]: DEBUG oslo_vmware.api [None req-2026e6a5-90f0-487b-95f0-c2b50f0a7e24 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385935, 'name': ReconfigVM_Task, 'duration_secs': 0.271519} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.819293] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-2026e6a5-90f0-487b-95f0-c2b50f0a7e24 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Reconfigured VM instance instance-00000062 to detach disk 2002 {{(pid=63202) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1105.824029] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0654ebee-f63a-496a-bc2f-14de25e299f1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.839127] env[63202]: DEBUG oslo_vmware.api [None req-2026e6a5-90f0-487b-95f0-c2b50f0a7e24 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Waiting for the task: (returnval){ [ 1105.839127] env[63202]: value = "task-1385936" [ 1105.839127] env[63202]: _type = "Task" [ 1105.839127] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.847914] env[63202]: DEBUG oslo_vmware.api [None req-2026e6a5-90f0-487b-95f0-c2b50f0a7e24 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385936, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.076425] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Releasing lock "refresh_cache-1cc050d2-1055-40b0-9769-dc03f7995da3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1106.130539] env[63202]: DEBUG oslo_concurrency.lockutils [None req-97fd197f-e5c0-4da9-8518-2059ebf5ab11 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquiring lock "8437f856-a707-49c5-b8eb-5a22cdb990f8" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1106.130799] env[63202]: DEBUG oslo_concurrency.lockutils [None req-97fd197f-e5c0-4da9-8518-2059ebf5ab11 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lock "8437f856-a707-49c5-b8eb-5a22cdb990f8" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1106.348623] env[63202]: DEBUG oslo_vmware.api [None req-2026e6a5-90f0-487b-95f0-c2b50f0a7e24 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385936, 'name': ReconfigVM_Task, 'duration_secs': 0.119992} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.348935] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-2026e6a5-90f0-487b-95f0-c2b50f0a7e24 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294228', 'volume_id': 'd200389d-4a94-4c63-8aa4-4d2b298bf88f', 'name': 'volume-d200389d-4a94-4c63-8aa4-4d2b298bf88f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e1e2e8a4-80aa-45eb-a90e-d20cba2943ee', 'attached_at': '', 'detached_at': '', 'volume_id': 'd200389d-4a94-4c63-8aa4-4d2b298bf88f', 'serial': 'd200389d-4a94-4c63-8aa4-4d2b298bf88f'} {{(pid=63202) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1106.601041] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0966a5d1-b2b1-4d79-afce-975ea99a8ab5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.619125] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e56c3bd9-510c-48fa-857e-b94d498bd62e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.625410] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Updating instance '1cc050d2-1055-40b0-9769-dc03f7995da3' progress to 83 {{(pid=63202) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1106.633782] env[63202]: INFO nova.compute.manager [None req-97fd197f-e5c0-4da9-8518-2059ebf5ab11 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Detaching volume 81b406ab-e721-4271-b298-75cded690f20 [ 1106.666101] env[63202]: INFO nova.virt.block_device [None req-97fd197f-e5c0-4da9-8518-2059ebf5ab11 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Attempting to driver detach volume 81b406ab-e721-4271-b298-75cded690f20 from mountpoint /dev/sdb [ 1106.666323] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-97fd197f-e5c0-4da9-8518-2059ebf5ab11 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Volume detach. Driver type: vmdk {{(pid=63202) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1106.666511] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-97fd197f-e5c0-4da9-8518-2059ebf5ab11 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294210', 'volume_id': '81b406ab-e721-4271-b298-75cded690f20', 'name': 'volume-81b406ab-e721-4271-b298-75cded690f20', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '8437f856-a707-49c5-b8eb-5a22cdb990f8', 'attached_at': '', 'detached_at': '', 'volume_id': '81b406ab-e721-4271-b298-75cded690f20', 'serial': '81b406ab-e721-4271-b298-75cded690f20'} {{(pid=63202) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1106.667376] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc54868a-f679-430f-91ca-ce56843b2706 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.689650] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd79c805-6839-4864-8361-6b7c59224477 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.695516] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57dd3931-bd41-4d75-af23-ed354850e729 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.715145] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e1c0c72-b5c8-4b24-8b56-4eb0ae05b54e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.731152] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-97fd197f-e5c0-4da9-8518-2059ebf5ab11 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] The volume has not been displaced from its original location: [datastore1] volume-81b406ab-e721-4271-b298-75cded690f20/volume-81b406ab-e721-4271-b298-75cded690f20.vmdk. No consolidation needed. {{(pid=63202) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1106.736372] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-97fd197f-e5c0-4da9-8518-2059ebf5ab11 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Reconfiguring VM instance instance-00000055 to detach disk 2001 {{(pid=63202) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1106.736655] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ee7e707a-6aa5-48be-ba8c-b4b12cb09d15 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.753991] env[63202]: DEBUG oslo_vmware.api [None req-97fd197f-e5c0-4da9-8518-2059ebf5ab11 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1106.753991] env[63202]: value = "task-1385937" [ 1106.753991] env[63202]: _type = "Task" [ 1106.753991] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.761465] env[63202]: DEBUG oslo_vmware.api [None req-97fd197f-e5c0-4da9-8518-2059ebf5ab11 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385937, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.891327] env[63202]: DEBUG nova.objects.instance [None req-2026e6a5-90f0-487b-95f0-c2b50f0a7e24 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lazy-loading 'flavor' on Instance uuid e1e2e8a4-80aa-45eb-a90e-d20cba2943ee {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1107.132036] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1107.132332] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c8984a8d-8f6b-4fb3-a7e1-d677f053005c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.138492] env[63202]: DEBUG oslo_vmware.api [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 1107.138492] env[63202]: value = "task-1385938" [ 1107.138492] env[63202]: _type = "Task" [ 1107.138492] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.145859] env[63202]: DEBUG oslo_vmware.api [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385938, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.263703] env[63202]: DEBUG oslo_vmware.api [None req-97fd197f-e5c0-4da9-8518-2059ebf5ab11 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385937, 'name': ReconfigVM_Task, 'duration_secs': 0.204049} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.263988] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-97fd197f-e5c0-4da9-8518-2059ebf5ab11 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Reconfigured VM instance instance-00000055 to detach disk 2001 {{(pid=63202) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1107.268743] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-17d38196-31a1-4bad-bbff-cea649a50417 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.283809] env[63202]: DEBUG oslo_vmware.api [None req-97fd197f-e5c0-4da9-8518-2059ebf5ab11 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1107.283809] env[63202]: value = "task-1385939" [ 1107.283809] env[63202]: _type = "Task" [ 1107.283809] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.294182] env[63202]: DEBUG oslo_vmware.api [None req-97fd197f-e5c0-4da9-8518-2059ebf5ab11 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385939, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.648998] env[63202]: DEBUG oslo_vmware.api [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385938, 'name': PowerOnVM_Task, 'duration_secs': 0.380389} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.649287] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1107.649474] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-2ca99cc8-03f4-4e3e-bc5a-8eef5722a56a tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Updating instance '1cc050d2-1055-40b0-9769-dc03f7995da3' progress to 100 {{(pid=63202) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1107.794371] env[63202]: DEBUG oslo_vmware.api [None req-97fd197f-e5c0-4da9-8518-2059ebf5ab11 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385939, 'name': ReconfigVM_Task, 'duration_secs': 0.143387} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.794722] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-97fd197f-e5c0-4da9-8518-2059ebf5ab11 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294210', 'volume_id': '81b406ab-e721-4271-b298-75cded690f20', 'name': 'volume-81b406ab-e721-4271-b298-75cded690f20', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '8437f856-a707-49c5-b8eb-5a22cdb990f8', 'attached_at': '', 'detached_at': '', 'volume_id': '81b406ab-e721-4271-b298-75cded690f20', 'serial': '81b406ab-e721-4271-b298-75cded690f20'} {{(pid=63202) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1107.899086] env[63202]: DEBUG oslo_concurrency.lockutils [None req-2026e6a5-90f0-487b-95f0-c2b50f0a7e24 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lock "e1e2e8a4-80aa-45eb-a90e-d20cba2943ee" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.208s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1108.337590] env[63202]: DEBUG nova.objects.instance [None req-97fd197f-e5c0-4da9-8518-2059ebf5ab11 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lazy-loading 'flavor' on Instance uuid 8437f856-a707-49c5-b8eb-5a22cdb990f8 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1109.072052] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b6c66fe0-532a-432b-862e-b24159b4a001 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Acquiring lock "e1e2e8a4-80aa-45eb-a90e-d20cba2943ee" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1109.072354] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b6c66fe0-532a-432b-862e-b24159b4a001 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lock "e1e2e8a4-80aa-45eb-a90e-d20cba2943ee" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1109.072631] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b6c66fe0-532a-432b-862e-b24159b4a001 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Acquiring lock "e1e2e8a4-80aa-45eb-a90e-d20cba2943ee-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1109.072760] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b6c66fe0-532a-432b-862e-b24159b4a001 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lock "e1e2e8a4-80aa-45eb-a90e-d20cba2943ee-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1109.072934] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b6c66fe0-532a-432b-862e-b24159b4a001 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lock "e1e2e8a4-80aa-45eb-a90e-d20cba2943ee-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1109.075259] env[63202]: INFO nova.compute.manager [None req-b6c66fe0-532a-432b-862e-b24159b4a001 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Terminating instance [ 1109.077168] env[63202]: DEBUG nova.compute.manager [None req-b6c66fe0-532a-432b-862e-b24159b4a001 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1109.077370] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-b6c66fe0-532a-432b-862e-b24159b4a001 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1109.078216] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd806ed5-30d8-4f89-b13f-9e3cc0c58bc8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.086648] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6c66fe0-532a-432b-862e-b24159b4a001 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1109.087980] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-614d8a73-da47-41e1-9ca9-4339a01b4679 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.097677] env[63202]: DEBUG oslo_vmware.api [None req-b6c66fe0-532a-432b-862e-b24159b4a001 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Waiting for the task: (returnval){ [ 1109.097677] env[63202]: value = "task-1385940" [ 1109.097677] env[63202]: _type = "Task" [ 1109.097677] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.107184] env[63202]: DEBUG oslo_vmware.api [None req-b6c66fe0-532a-432b-862e-b24159b4a001 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385940, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.345130] env[63202]: DEBUG oslo_concurrency.lockutils [None req-97fd197f-e5c0-4da9-8518-2059ebf5ab11 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lock "8437f856-a707-49c5-b8eb-5a22cdb990f8" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.214s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1109.594172] env[63202]: DEBUG oslo_concurrency.lockutils [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "1cc050d2-1055-40b0-9769-dc03f7995da3" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1109.594486] env[63202]: DEBUG oslo_concurrency.lockutils [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "1cc050d2-1055-40b0-9769-dc03f7995da3" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1109.594698] env[63202]: DEBUG nova.compute.manager [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Going to confirm migration 3 {{(pid=63202) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1109.607864] env[63202]: DEBUG oslo_vmware.api [None req-b6c66fe0-532a-432b-862e-b24159b4a001 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385940, 'name': PowerOffVM_Task, 'duration_secs': 0.198302} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.608533] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6c66fe0-532a-432b-862e-b24159b4a001 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1109.608533] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-b6c66fe0-532a-432b-862e-b24159b4a001 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1109.608663] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e8da53c0-c171-4f08-9578-81575dfc9ad1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.680801] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-b6c66fe0-532a-432b-862e-b24159b4a001 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1109.680801] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-b6c66fe0-532a-432b-862e-b24159b4a001 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Deleting contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1109.680801] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-b6c66fe0-532a-432b-862e-b24159b4a001 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Deleting the datastore file [datastore1] e1e2e8a4-80aa-45eb-a90e-d20cba2943ee {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1109.681078] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7cd93d7a-220c-4f22-ab2f-71cecbb0bee2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.687984] env[63202]: DEBUG oslo_vmware.api [None req-b6c66fe0-532a-432b-862e-b24159b4a001 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Waiting for the task: (returnval){ [ 1109.687984] env[63202]: value = "task-1385942" [ 1109.687984] env[63202]: _type = "Task" [ 1109.687984] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.696279] env[63202]: DEBUG oslo_vmware.api [None req-b6c66fe0-532a-432b-862e-b24159b4a001 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385942, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.159762] env[63202]: DEBUG oslo_concurrency.lockutils [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "refresh_cache-1cc050d2-1055-40b0-9769-dc03f7995da3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1110.159941] env[63202]: DEBUG oslo_concurrency.lockutils [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquired lock "refresh_cache-1cc050d2-1055-40b0-9769-dc03f7995da3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1110.160174] env[63202]: DEBUG nova.network.neutron [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1110.160381] env[63202]: DEBUG nova.objects.instance [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lazy-loading 'info_cache' on Instance uuid 1cc050d2-1055-40b0-9769-dc03f7995da3 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1110.197988] env[63202]: DEBUG oslo_vmware.api [None req-b6c66fe0-532a-432b-862e-b24159b4a001 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Task: {'id': task-1385942, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145963} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.198270] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-b6c66fe0-532a-432b-862e-b24159b4a001 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1110.198457] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-b6c66fe0-532a-432b-862e-b24159b4a001 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Deleted contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1110.198636] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-b6c66fe0-532a-432b-862e-b24159b4a001 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1110.198810] env[63202]: INFO nova.compute.manager [None req-b6c66fe0-532a-432b-862e-b24159b4a001 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1110.199068] env[63202]: DEBUG oslo.service.loopingcall [None req-b6c66fe0-532a-432b-862e-b24159b4a001 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1110.199272] env[63202]: DEBUG nova.compute.manager [-] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1110.199365] env[63202]: DEBUG nova.network.neutron [-] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1110.582619] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8f26d3d8-dd68-4a94-96dc-5b623f01be3d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquiring lock "8437f856-a707-49c5-b8eb-5a22cdb990f8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1110.582972] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8f26d3d8-dd68-4a94-96dc-5b623f01be3d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lock "8437f856-a707-49c5-b8eb-5a22cdb990f8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1110.583138] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8f26d3d8-dd68-4a94-96dc-5b623f01be3d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquiring lock "8437f856-a707-49c5-b8eb-5a22cdb990f8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1110.583339] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8f26d3d8-dd68-4a94-96dc-5b623f01be3d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lock "8437f856-a707-49c5-b8eb-5a22cdb990f8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1110.583513] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8f26d3d8-dd68-4a94-96dc-5b623f01be3d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lock "8437f856-a707-49c5-b8eb-5a22cdb990f8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1110.586043] env[63202]: INFO nova.compute.manager [None req-8f26d3d8-dd68-4a94-96dc-5b623f01be3d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Terminating instance [ 1110.587817] env[63202]: DEBUG nova.compute.manager [None req-8f26d3d8-dd68-4a94-96dc-5b623f01be3d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1110.588018] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-8f26d3d8-dd68-4a94-96dc-5b623f01be3d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1110.588934] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a628f6dc-9e2f-4684-8e2b-4ee08eab3830 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.596367] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f26d3d8-dd68-4a94-96dc-5b623f01be3d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1110.596605] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dbf7be11-38cf-4097-9f20-045c71abd53d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.605411] env[63202]: DEBUG oslo_vmware.api [None req-8f26d3d8-dd68-4a94-96dc-5b623f01be3d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1110.605411] env[63202]: value = "task-1385943" [ 1110.605411] env[63202]: _type = "Task" [ 1110.605411] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.612755] env[63202]: DEBUG oslo_vmware.api [None req-8f26d3d8-dd68-4a94-96dc-5b623f01be3d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385943, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.718373] env[63202]: DEBUG nova.compute.manager [req-1b9db746-c3a9-4575-bd8a-5d33d5328bcb req-37d062a2-f6b6-4cc6-a3b4-48c4520bfb90 service nova] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Received event network-vif-deleted-3cec3f36-0fb8-45c0-be6e-48cd6cf58aa3 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1110.718524] env[63202]: INFO nova.compute.manager [req-1b9db746-c3a9-4575-bd8a-5d33d5328bcb req-37d062a2-f6b6-4cc6-a3b4-48c4520bfb90 service nova] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Neutron deleted interface 3cec3f36-0fb8-45c0-be6e-48cd6cf58aa3; detaching it from the instance and deleting it from the info cache [ 1110.718779] env[63202]: DEBUG nova.network.neutron [req-1b9db746-c3a9-4575-bd8a-5d33d5328bcb req-37d062a2-f6b6-4cc6-a3b4-48c4520bfb90 service nova] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1111.115129] env[63202]: DEBUG oslo_vmware.api [None req-8f26d3d8-dd68-4a94-96dc-5b623f01be3d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385943, 'name': PowerOffVM_Task, 'duration_secs': 0.18006} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.115385] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f26d3d8-dd68-4a94-96dc-5b623f01be3d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1111.115552] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-8f26d3d8-dd68-4a94-96dc-5b623f01be3d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1111.115797] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ac0ed3d2-f0cc-47d9-9907-10edfb350cf3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.159947] env[63202]: DEBUG nova.network.neutron [-] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1111.178847] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-8f26d3d8-dd68-4a94-96dc-5b623f01be3d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1111.179086] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-8f26d3d8-dd68-4a94-96dc-5b623f01be3d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Deleting contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1111.179278] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f26d3d8-dd68-4a94-96dc-5b623f01be3d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Deleting the datastore file [datastore1] 8437f856-a707-49c5-b8eb-5a22cdb990f8 {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1111.179534] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4eb01e6c-7698-4190-82f4-4cd9b6b36726 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.185868] env[63202]: DEBUG oslo_vmware.api [None req-8f26d3d8-dd68-4a94-96dc-5b623f01be3d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1111.185868] env[63202]: value = "task-1385945" [ 1111.185868] env[63202]: _type = "Task" [ 1111.185868] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.194529] env[63202]: DEBUG oslo_vmware.api [None req-8f26d3d8-dd68-4a94-96dc-5b623f01be3d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385945, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.221086] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-06591b5f-1b0d-4136-8e25-c27afc39ebf7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.229817] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5847f365-7691-471b-8eba-8c14362cccbd {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.258603] env[63202]: DEBUG nova.compute.manager [req-1b9db746-c3a9-4575-bd8a-5d33d5328bcb req-37d062a2-f6b6-4cc6-a3b4-48c4520bfb90 service nova] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Detach interface failed, port_id=3cec3f36-0fb8-45c0-be6e-48cd6cf58aa3, reason: Instance e1e2e8a4-80aa-45eb-a90e-d20cba2943ee could not be found. {{(pid=63202) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1111.390953] env[63202]: DEBUG nova.network.neutron [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Updating instance_info_cache with network_info: [{"id": "ad8d36fb-afb1-480f-815c-6d60c7758b42", "address": "fa:16:3e:47:0d:32", "network": {"id": "5a3b4c9b-2ca7-4f8b-8bbb-ea000db91402", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-543677750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f508ec04edc844a19640a8a85f27e5b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad8d36fb-af", "ovs_interfaceid": "ad8d36fb-afb1-480f-815c-6d60c7758b42", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1111.663800] env[63202]: INFO nova.compute.manager [-] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Took 1.46 seconds to deallocate network for instance. [ 1111.695353] env[63202]: DEBUG oslo_vmware.api [None req-8f26d3d8-dd68-4a94-96dc-5b623f01be3d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385945, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.124997} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.696906] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f26d3d8-dd68-4a94-96dc-5b623f01be3d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1111.696906] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-8f26d3d8-dd68-4a94-96dc-5b623f01be3d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Deleted contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1111.696906] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-8f26d3d8-dd68-4a94-96dc-5b623f01be3d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1111.696906] env[63202]: INFO nova.compute.manager [None req-8f26d3d8-dd68-4a94-96dc-5b623f01be3d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1111.696906] env[63202]: DEBUG oslo.service.loopingcall [None req-8f26d3d8-dd68-4a94-96dc-5b623f01be3d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1111.697143] env[63202]: DEBUG nova.compute.manager [-] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1111.697143] env[63202]: DEBUG nova.network.neutron [-] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1111.896746] env[63202]: DEBUG oslo_concurrency.lockutils [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Releasing lock "refresh_cache-1cc050d2-1055-40b0-9769-dc03f7995da3" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1111.896746] env[63202]: DEBUG nova.objects.instance [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lazy-loading 'migration_context' on Instance uuid 1cc050d2-1055-40b0-9769-dc03f7995da3 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1112.170469] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b6c66fe0-532a-432b-862e-b24159b4a001 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1112.170753] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b6c66fe0-532a-432b-862e-b24159b4a001 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1112.170976] env[63202]: DEBUG nova.objects.instance [None req-b6c66fe0-532a-432b-862e-b24159b4a001 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lazy-loading 'resources' on Instance uuid e1e2e8a4-80aa-45eb-a90e-d20cba2943ee {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1112.399186] env[63202]: DEBUG nova.objects.base [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Object Instance<1cc050d2-1055-40b0-9769-dc03f7995da3> lazy-loaded attributes: info_cache,migration_context {{(pid=63202) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1112.400116] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-668f65da-1b24-4fe0-9fe5-998acef7efe3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.420767] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-abc7658f-4d12-4349-99fa-c400b7a81047 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.426351] env[63202]: DEBUG oslo_vmware.api [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 1112.426351] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]522a7a84-991a-7cd2-dd1c-f221a90e282a" [ 1112.426351] env[63202]: _type = "Task" [ 1112.426351] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.434180] env[63202]: DEBUG oslo_vmware.api [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]522a7a84-991a-7cd2-dd1c-f221a90e282a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.630531] env[63202]: DEBUG nova.network.neutron [-] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1112.762926] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97c31e80-d7d9-44de-9180-9038f34848c5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.770475] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0352267b-d02e-487a-aa4a-86d4c2941ec8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.799963] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e410f96-3e45-4971-aa68-8beccc23420e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.805278] env[63202]: DEBUG nova.compute.manager [req-bc333008-3dcb-4264-976e-7846b193f085 req-ed21a5ab-ff4d-45ce-9334-bde162b1450f service nova] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Received event network-vif-deleted-07cc827a-876a-4432-a5fc-9ba8920dc5f9 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1112.809027] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c0b18d1-26e9-434f-b7b7-6c3edb7d0db8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.823469] env[63202]: DEBUG nova.compute.provider_tree [None req-b6c66fe0-532a-432b-862e-b24159b4a001 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1112.936619] env[63202]: DEBUG oslo_vmware.api [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]522a7a84-991a-7cd2-dd1c-f221a90e282a, 'name': SearchDatastore_Task, 'duration_secs': 0.009858} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.936920] env[63202]: DEBUG oslo_concurrency.lockutils [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.133717] env[63202]: INFO nova.compute.manager [-] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Took 1.44 seconds to deallocate network for instance. [ 1113.326680] env[63202]: DEBUG nova.scheduler.client.report [None req-b6c66fe0-532a-432b-862e-b24159b4a001 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1113.640852] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8f26d3d8-dd68-4a94-96dc-5b623f01be3d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.832233] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b6c66fe0-532a-432b-862e-b24159b4a001 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.661s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1113.835486] env[63202]: DEBUG oslo_concurrency.lockutils [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.898s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.857224] env[63202]: INFO nova.scheduler.client.report [None req-b6c66fe0-532a-432b-862e-b24159b4a001 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Deleted allocations for instance e1e2e8a4-80aa-45eb-a90e-d20cba2943ee [ 1114.114579] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Acquiring lock "9646810e-06d2-4de0-8f0c-e8719541da53" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1114.114845] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Lock "9646810e-06d2-4de0-8f0c-e8719541da53" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1114.115037] env[63202]: INFO nova.compute.manager [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Shelving [ 1114.364687] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b6c66fe0-532a-432b-862e-b24159b4a001 tempest-AttachVolumeTestJSON-2000870146 tempest-AttachVolumeTestJSON-2000870146-project-member] Lock "e1e2e8a4-80aa-45eb-a90e-d20cba2943ee" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.292s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1114.410964] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2364e454-8656-4577-8c73-56e231770aa5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.419038] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c439765-0c76-4a86-a114-bc6241d65d57 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.454208] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a37263ba-7661-43b0-af4e-83cd5c64932f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.462724] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3a30d9f-e8eb-4688-bf38-0d68af56a9ec {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.477042] env[63202]: DEBUG nova.compute.provider_tree [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1114.622121] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1114.622386] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-05c8c31a-ac0b-4f87-9f42-d41ff6506823 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.630611] env[63202]: DEBUG oslo_vmware.api [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Waiting for the task: (returnval){ [ 1114.630611] env[63202]: value = "task-1385946" [ 1114.630611] env[63202]: _type = "Task" [ 1114.630611] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.638809] env[63202]: DEBUG oslo_vmware.api [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385946, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.981017] env[63202]: DEBUG nova.scheduler.client.report [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1115.140799] env[63202]: DEBUG oslo_vmware.api [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385946, 'name': PowerOffVM_Task, 'duration_secs': 0.210544} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.141089] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1115.141836] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-864f3bef-2805-4da4-a7f8-0b1f949e38c4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.159438] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7d6564e-c3a5-411f-aac4-6d3e64357ca3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.669021] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Creating Snapshot of the VM instance {{(pid=63202) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1115.669758] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-0fa13282-28f2-4f58-acde-10f8476c50eb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.677620] env[63202]: DEBUG oslo_vmware.api [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Waiting for the task: (returnval){ [ 1115.677620] env[63202]: value = "task-1385948" [ 1115.677620] env[63202]: _type = "Task" [ 1115.677620] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.685248] env[63202]: DEBUG oslo_vmware.api [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385948, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.991932] env[63202]: DEBUG oslo_concurrency.lockutils [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.156s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1115.995036] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8f26d3d8-dd68-4a94-96dc-5b623f01be3d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.354s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1115.995218] env[63202]: DEBUG nova.objects.instance [None req-8f26d3d8-dd68-4a94-96dc-5b623f01be3d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lazy-loading 'resources' on Instance uuid 8437f856-a707-49c5-b8eb-5a22cdb990f8 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1116.188350] env[63202]: DEBUG oslo_vmware.api [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385948, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.558874] env[63202]: INFO nova.scheduler.client.report [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Deleted allocation for migration d0fa269f-50a0-48dc-b06a-d3d25ba42c25 [ 1116.566544] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-019a22b6-f18c-4a30-b242-33f5e7ea7970 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.574619] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06e34490-bec7-40ce-9d33-a82e51dfbacd {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.604889] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-367cf947-2e21-4005-8409-d8e25528ae8c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.612643] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8be5a7bc-1041-4216-a1b5-afdeb36bbb5b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.625804] env[63202]: DEBUG nova.compute.provider_tree [None req-8f26d3d8-dd68-4a94-96dc-5b623f01be3d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1116.687710] env[63202]: DEBUG oslo_vmware.api [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385948, 'name': CreateSnapshot_Task, 'duration_secs': 0.558379} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.687971] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Created Snapshot of the VM instance {{(pid=63202) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1116.688697] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d17a0e06-4629-4798-a0a3-e0f497b973da {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.063986] env[63202]: DEBUG oslo_concurrency.lockutils [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "1cc050d2-1055-40b0-9769-dc03f7995da3" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.469s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1117.129107] env[63202]: DEBUG nova.scheduler.client.report [None req-8f26d3d8-dd68-4a94-96dc-5b623f01be3d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1117.205801] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Creating linked-clone VM from snapshot {{(pid=63202) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1117.206106] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-50a2f6d1-5bff-4767-9f05-342ed0d74793 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.214815] env[63202]: DEBUG oslo_vmware.api [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Waiting for the task: (returnval){ [ 1117.214815] env[63202]: value = "task-1385950" [ 1117.214815] env[63202]: _type = "Task" [ 1117.214815] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.222323] env[63202]: DEBUG oslo_vmware.api [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385950, 'name': CloneVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.633697] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8f26d3d8-dd68-4a94-96dc-5b623f01be3d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.639s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1117.648990] env[63202]: INFO nova.scheduler.client.report [None req-8f26d3d8-dd68-4a94-96dc-5b623f01be3d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Deleted allocations for instance 8437f856-a707-49c5-b8eb-5a22cdb990f8 [ 1117.724315] env[63202]: DEBUG oslo_vmware.api [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385950, 'name': CloneVM_Task} progress is 94%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.117416] env[63202]: DEBUG oslo_concurrency.lockutils [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "1cc050d2-1055-40b0-9769-dc03f7995da3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1118.117416] env[63202]: DEBUG oslo_concurrency.lockutils [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "1cc050d2-1055-40b0-9769-dc03f7995da3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1118.117416] env[63202]: DEBUG oslo_concurrency.lockutils [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "1cc050d2-1055-40b0-9769-dc03f7995da3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1118.117416] env[63202]: DEBUG oslo_concurrency.lockutils [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "1cc050d2-1055-40b0-9769-dc03f7995da3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1118.117852] env[63202]: DEBUG oslo_concurrency.lockutils [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "1cc050d2-1055-40b0-9769-dc03f7995da3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1118.118533] env[63202]: INFO nova.compute.manager [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Terminating instance [ 1118.120609] env[63202]: DEBUG nova.compute.manager [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1118.120932] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1118.121879] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9eee166-8169-404c-b877-bcb42dc1f1f8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.130027] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1118.130027] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-92d7fad4-cefb-471b-a462-6cc26bad1ec5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.135535] env[63202]: DEBUG oslo_vmware.api [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 1118.135535] env[63202]: value = "task-1385951" [ 1118.135535] env[63202]: _type = "Task" [ 1118.135535] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.143655] env[63202]: DEBUG oslo_vmware.api [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385951, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.155890] env[63202]: DEBUG oslo_concurrency.lockutils [None req-8f26d3d8-dd68-4a94-96dc-5b623f01be3d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lock "8437f856-a707-49c5-b8eb-5a22cdb990f8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.573s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1118.226190] env[63202]: DEBUG oslo_vmware.api [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385950, 'name': CloneVM_Task} progress is 100%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.645706] env[63202]: DEBUG oslo_vmware.api [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385951, 'name': PowerOffVM_Task, 'duration_secs': 0.202049} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.645996] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1118.646177] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1118.646427] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7d175e3e-b169-4194-b7e0-800bee28a009 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.709151] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1118.709378] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Deleting contents of the VM from datastore datastore2 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1118.709578] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Deleting the datastore file [datastore2] 1cc050d2-1055-40b0-9769-dc03f7995da3 {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1118.709828] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cb18c23c-61d6-4606-80df-c9f47f60648d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.715887] env[63202]: DEBUG oslo_vmware.api [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for the task: (returnval){ [ 1118.715887] env[63202]: value = "task-1385953" [ 1118.715887] env[63202]: _type = "Task" [ 1118.715887] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.727180] env[63202]: DEBUG oslo_vmware.api [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385950, 'name': CloneVM_Task, 'duration_secs': 1.050244} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.730153] env[63202]: INFO nova.virt.vmwareapi.vmops [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Created linked-clone VM from snapshot [ 1118.730341] env[63202]: DEBUG oslo_vmware.api [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385953, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.731029] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b772becb-f020-40d4-9f22-a2c4eaa2c65f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.739070] env[63202]: DEBUG nova.virt.vmwareapi.images [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Uploading image 06831fb6-0dd7-4e39-a9d1-f69598985fd2 {{(pid=63202) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1118.763877] env[63202]: DEBUG oslo_vmware.rw_handles [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1118.763877] env[63202]: value = "vm-294230" [ 1118.763877] env[63202]: _type = "VirtualMachine" [ 1118.763877] env[63202]: }. {{(pid=63202) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1118.764174] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-52875a60-fbdb-4646-a198-a8f6cbf68ffb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.770827] env[63202]: DEBUG oslo_vmware.rw_handles [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Lease: (returnval){ [ 1118.770827] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]5278852a-3088-fee0-ae4b-554a2752baa8" [ 1118.770827] env[63202]: _type = "HttpNfcLease" [ 1118.770827] env[63202]: } obtained for exporting VM: (result){ [ 1118.770827] env[63202]: value = "vm-294230" [ 1118.770827] env[63202]: _type = "VirtualMachine" [ 1118.770827] env[63202]: }. {{(pid=63202) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1118.771230] env[63202]: DEBUG oslo_vmware.api [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Waiting for the lease: (returnval){ [ 1118.771230] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]5278852a-3088-fee0-ae4b-554a2752baa8" [ 1118.771230] env[63202]: _type = "HttpNfcLease" [ 1118.771230] env[63202]: } to be ready. {{(pid=63202) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1118.777014] env[63202]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1118.777014] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]5278852a-3088-fee0-ae4b-554a2752baa8" [ 1118.777014] env[63202]: _type = "HttpNfcLease" [ 1118.777014] env[63202]: } is initializing. {{(pid=63202) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1119.229682] env[63202]: DEBUG oslo_vmware.api [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Task: {'id': task-1385953, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139095} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.229964] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1119.230155] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Deleted contents of the VM from datastore datastore2 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1119.230371] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1119.230552] env[63202]: INFO nova.compute.manager [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1119.230794] env[63202]: DEBUG oslo.service.loopingcall [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1119.230986] env[63202]: DEBUG nova.compute.manager [-] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1119.231091] env[63202]: DEBUG nova.network.neutron [-] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1119.279368] env[63202]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1119.279368] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]5278852a-3088-fee0-ae4b-554a2752baa8" [ 1119.279368] env[63202]: _type = "HttpNfcLease" [ 1119.279368] env[63202]: } is ready. {{(pid=63202) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1119.279709] env[63202]: DEBUG oslo_vmware.rw_handles [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1119.279709] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]5278852a-3088-fee0-ae4b-554a2752baa8" [ 1119.279709] env[63202]: _type = "HttpNfcLease" [ 1119.279709] env[63202]: }. {{(pid=63202) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1119.280463] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0560224c-930c-4a9a-ad52-34d6ac480212 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.287930] env[63202]: DEBUG oslo_vmware.rw_handles [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c2dab5-f02a-35e2-86af-a9a34c9f8fd6/disk-0.vmdk from lease info. {{(pid=63202) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1119.288126] env[63202]: DEBUG oslo_vmware.rw_handles [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c2dab5-f02a-35e2-86af-a9a34c9f8fd6/disk-0.vmdk for reading. {{(pid=63202) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1119.380029] env[63202]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-d37894e5-050e-408e-84f0-a022bda6b125 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.553443] env[63202]: DEBUG nova.compute.manager [req-634ffa38-3f7a-4a84-b5f5-98213a0c11cd req-b4941a99-4788-492f-b17f-a4c5c7bc4eb3 service nova] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Received event network-vif-deleted-ad8d36fb-afb1-480f-815c-6d60c7758b42 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1119.553721] env[63202]: INFO nova.compute.manager [req-634ffa38-3f7a-4a84-b5f5-98213a0c11cd req-b4941a99-4788-492f-b17f-a4c5c7bc4eb3 service nova] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Neutron deleted interface ad8d36fb-afb1-480f-815c-6d60c7758b42; detaching it from the instance and deleting it from the info cache [ 1119.553818] env[63202]: DEBUG nova.network.neutron [req-634ffa38-3f7a-4a84-b5f5-98213a0c11cd req-b4941a99-4788-492f-b17f-a4c5c7bc4eb3 service nova] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1119.835847] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquiring lock "ed60fb5e-b08f-4846-942d-c4edde39d1ae" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1119.836152] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lock "ed60fb5e-b08f-4846-942d-c4edde39d1ae" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1120.028096] env[63202]: DEBUG nova.network.neutron [-] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1120.057114] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1aba40ef-387c-4b92-85b1-9ff16c249423 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.066895] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-944e33f8-bf02-44f2-b018-4c12a8df0c53 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.095206] env[63202]: DEBUG nova.compute.manager [req-634ffa38-3f7a-4a84-b5f5-98213a0c11cd req-b4941a99-4788-492f-b17f-a4c5c7bc4eb3 service nova] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Detach interface failed, port_id=ad8d36fb-afb1-480f-815c-6d60c7758b42, reason: Instance 1cc050d2-1055-40b0-9769-dc03f7995da3 could not be found. {{(pid=63202) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1120.339639] env[63202]: DEBUG nova.compute.manager [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1120.533250] env[63202]: INFO nova.compute.manager [-] [instance: 1cc050d2-1055-40b0-9769-dc03f7995da3] Took 1.30 seconds to deallocate network for instance. [ 1120.862351] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1120.863362] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1120.864442] env[63202]: INFO nova.compute.claims [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1121.040169] env[63202]: DEBUG oslo_concurrency.lockutils [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1121.524035] env[63202]: DEBUG oslo_concurrency.lockutils [None req-250db1ba-9f5e-4459-9667-6d7651d6c7a0 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquiring lock "214c63b6-35e3-46e1-a9de-f3cea251c3c9" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1121.524385] env[63202]: DEBUG oslo_concurrency.lockutils [None req-250db1ba-9f5e-4459-9667-6d7651d6c7a0 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "214c63b6-35e3-46e1-a9de-f3cea251c3c9" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1121.933943] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0095f0fe-0c60-4ad3-8121-07d82dbf3670 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.941737] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-085adbbb-8562-40cc-b730-592664b78161 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.970633] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77f339f7-ea60-459f-8a93-96ba94644345 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.977892] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fdf9728-d5f5-4031-9844-5f0f5273674c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.990864] env[63202]: DEBUG nova.compute.provider_tree [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1122.027327] env[63202]: DEBUG nova.compute.utils [None req-250db1ba-9f5e-4459-9667-6d7651d6c7a0 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1122.253900] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1122.254092] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1122.254274] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1122.254414] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Cleaning up deleted instances {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 1122.494241] env[63202]: DEBUG nova.scheduler.client.report [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1122.530428] env[63202]: DEBUG oslo_concurrency.lockutils [None req-250db1ba-9f5e-4459-9667-6d7651d6c7a0 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "214c63b6-35e3-46e1-a9de-f3cea251c3c9" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1122.766557] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] There are 40 instances to clean {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 1122.766842] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: bed3be83-e467-4466-ad1f-9ca89bfa87a8] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1122.999027] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.136s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1122.999448] env[63202]: DEBUG nova.compute.manager [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1123.002148] env[63202]: DEBUG oslo_concurrency.lockutils [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.962s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.002392] env[63202]: DEBUG oslo_concurrency.lockutils [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.023877] env[63202]: INFO nova.scheduler.client.report [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Deleted allocations for instance 1cc050d2-1055-40b0-9769-dc03f7995da3 [ 1123.270803] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 6e172a2a-1918-430c-84ce-005fef97bbf1] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1123.506508] env[63202]: DEBUG nova.compute.utils [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1123.508052] env[63202]: DEBUG nova.compute.manager [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1123.508270] env[63202]: DEBUG nova.network.neutron [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1123.532134] env[63202]: DEBUG oslo_concurrency.lockutils [None req-340201f0-3dcb-453a-ae99-8550e9868734 tempest-DeleteServersTestJSON-1267463819 tempest-DeleteServersTestJSON-1267463819-project-member] Lock "1cc050d2-1055-40b0-9769-dc03f7995da3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.417s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.581566] env[63202]: DEBUG nova.policy [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3e9b38cd253441939b048166110aef47', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2b2de35030a9484094e964ffc30a822d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 1123.613720] env[63202]: DEBUG oslo_concurrency.lockutils [None req-250db1ba-9f5e-4459-9667-6d7651d6c7a0 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquiring lock "214c63b6-35e3-46e1-a9de-f3cea251c3c9" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.614057] env[63202]: DEBUG oslo_concurrency.lockutils [None req-250db1ba-9f5e-4459-9667-6d7651d6c7a0 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "214c63b6-35e3-46e1-a9de-f3cea251c3c9" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.614373] env[63202]: INFO nova.compute.manager [None req-250db1ba-9f5e-4459-9667-6d7651d6c7a0 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Attaching volume 71cb14b0-2665-4117-88e9-180e83bde620 to /dev/sdb [ 1123.652036] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-912b8916-5f99-4b05-b517-d86327ca80c0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.659190] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fc3321d-0c8c-4869-af66-1ac6916c595d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.671903] env[63202]: DEBUG nova.virt.block_device [None req-250db1ba-9f5e-4459-9667-6d7651d6c7a0 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Updating existing volume attachment record: df9d4c2a-29d1-4116-805f-34f6c5c029dd {{(pid=63202) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1123.774068] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 439ee94c-04f2-45d3-a486-81a216c8db4e] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1124.012064] env[63202]: DEBUG nova.compute.manager [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1124.072155] env[63202]: DEBUG nova.network.neutron [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Successfully created port: 0193eadb-322b-4278-b259-d56b23555561 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1124.277457] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 1123a12e-5218-415d-b286-2f005fe57b29] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1124.788240] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: e1e2e8a4-80aa-45eb-a90e-d20cba2943ee] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1125.026111] env[63202]: DEBUG nova.compute.manager [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1125.051651] env[63202]: DEBUG nova.virt.hardware [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1125.051923] env[63202]: DEBUG nova.virt.hardware [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1125.052097] env[63202]: DEBUG nova.virt.hardware [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1125.052323] env[63202]: DEBUG nova.virt.hardware [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1125.052493] env[63202]: DEBUG nova.virt.hardware [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1125.052655] env[63202]: DEBUG nova.virt.hardware [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1125.052951] env[63202]: DEBUG nova.virt.hardware [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1125.053020] env[63202]: DEBUG nova.virt.hardware [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1125.053205] env[63202]: DEBUG nova.virt.hardware [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1125.053402] env[63202]: DEBUG nova.virt.hardware [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1125.053576] env[63202]: DEBUG nova.virt.hardware [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1125.054509] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19b31bf9-27c0-4773-9ae4-26bff34e7190 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.063424] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c289fbf-4191-4c60-b58b-361b4acd875e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.294660] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 3a65ddea-2d3d-45a1-a175-c8fad756793d] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1125.578761] env[63202]: DEBUG nova.compute.manager [req-d08fe3c6-21c8-4db9-a03c-a093a99aaa26 req-1d4aae4d-7654-43b3-adad-ab9cc30f372f service nova] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Received event network-vif-plugged-0193eadb-322b-4278-b259-d56b23555561 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1125.578817] env[63202]: DEBUG oslo_concurrency.lockutils [req-d08fe3c6-21c8-4db9-a03c-a093a99aaa26 req-1d4aae4d-7654-43b3-adad-ab9cc30f372f service nova] Acquiring lock "ed60fb5e-b08f-4846-942d-c4edde39d1ae-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1125.578995] env[63202]: DEBUG oslo_concurrency.lockutils [req-d08fe3c6-21c8-4db9-a03c-a093a99aaa26 req-1d4aae4d-7654-43b3-adad-ab9cc30f372f service nova] Lock "ed60fb5e-b08f-4846-942d-c4edde39d1ae-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1125.579374] env[63202]: DEBUG oslo_concurrency.lockutils [req-d08fe3c6-21c8-4db9-a03c-a093a99aaa26 req-1d4aae4d-7654-43b3-adad-ab9cc30f372f service nova] Lock "ed60fb5e-b08f-4846-942d-c4edde39d1ae-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1125.579561] env[63202]: DEBUG nova.compute.manager [req-d08fe3c6-21c8-4db9-a03c-a093a99aaa26 req-1d4aae4d-7654-43b3-adad-ab9cc30f372f service nova] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] No waiting events found dispatching network-vif-plugged-0193eadb-322b-4278-b259-d56b23555561 {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1125.579750] env[63202]: WARNING nova.compute.manager [req-d08fe3c6-21c8-4db9-a03c-a093a99aaa26 req-1d4aae4d-7654-43b3-adad-ab9cc30f372f service nova] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Received unexpected event network-vif-plugged-0193eadb-322b-4278-b259-d56b23555561 for instance with vm_state building and task_state spawning. [ 1125.716698] env[63202]: DEBUG nova.network.neutron [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Successfully updated port: 0193eadb-322b-4278-b259-d56b23555561 {{(pid=63202) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1125.798554] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: c2b6f3a4-7414-4fc4-893b-9b613fc5e381] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1126.219147] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquiring lock "refresh_cache-ed60fb5e-b08f-4846-942d-c4edde39d1ae" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1126.219309] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquired lock "refresh_cache-ed60fb5e-b08f-4846-942d-c4edde39d1ae" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1126.219459] env[63202]: DEBUG nova.network.neutron [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1126.302265] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 7ac2bfb5-e8d0-4dbf-a647-7c3d65c9327b] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1126.767044] env[63202]: DEBUG nova.network.neutron [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1126.806454] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 3029c06f-4687-46f7-958b-831e4d0b241b] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1126.841239] env[63202]: DEBUG oslo_vmware.rw_handles [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c2dab5-f02a-35e2-86af-a9a34c9f8fd6/disk-0.vmdk. {{(pid=63202) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1126.842433] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e54ced7e-4df8-4f9d-8b59-da2d8258275d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.849244] env[63202]: DEBUG oslo_vmware.rw_handles [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c2dab5-f02a-35e2-86af-a9a34c9f8fd6/disk-0.vmdk is in state: ready. {{(pid=63202) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1126.849417] env[63202]: ERROR oslo_vmware.rw_handles [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c2dab5-f02a-35e2-86af-a9a34c9f8fd6/disk-0.vmdk due to incomplete transfer. [ 1126.849639] env[63202]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-c8f08c38-0051-4c40-826d-7735da975c07 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.856526] env[63202]: DEBUG oslo_vmware.rw_handles [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c2dab5-f02a-35e2-86af-a9a34c9f8fd6/disk-0.vmdk. {{(pid=63202) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1126.856727] env[63202]: DEBUG nova.virt.vmwareapi.images [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Uploaded image 06831fb6-0dd7-4e39-a9d1-f69598985fd2 to the Glance image server {{(pid=63202) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1126.859251] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Destroying the VM {{(pid=63202) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1126.859520] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-695bbebf-172d-4b05-9477-71678125d409 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.864759] env[63202]: DEBUG oslo_vmware.api [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Waiting for the task: (returnval){ [ 1126.864759] env[63202]: value = "task-1385959" [ 1126.864759] env[63202]: _type = "Task" [ 1126.864759] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.875123] env[63202]: DEBUG oslo_vmware.api [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385959, 'name': Destroy_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.157434] env[63202]: DEBUG nova.network.neutron [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Updating instance_info_cache with network_info: [{"id": "0193eadb-322b-4278-b259-d56b23555561", "address": "fa:16:3e:1e:87:93", "network": {"id": "06ab5813-9ad9-4021-9bdb-f2f02af8d73f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1714653503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b2de35030a9484094e964ffc30a822d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3093647a-bab7-4562-ada0-428725e8c0fc", "external-id": "nsx-vlan-transportzone-660", "segmentation_id": 660, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0193eadb-32", "ovs_interfaceid": "0193eadb-322b-4278-b259-d56b23555561", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1127.309745] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: b29a8043-2006-4255-8036-669d2be594d3] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1127.374977] env[63202]: DEBUG oslo_vmware.api [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385959, 'name': Destroy_Task, 'duration_secs': 0.325707} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.375267] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Destroyed the VM [ 1127.375506] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Deleting Snapshot of the VM instance {{(pid=63202) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1127.375758] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-93207a60-2b20-4058-9967-1be54b329b1e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.381651] env[63202]: DEBUG oslo_vmware.api [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Waiting for the task: (returnval){ [ 1127.381651] env[63202]: value = "task-1385960" [ 1127.381651] env[63202]: _type = "Task" [ 1127.381651] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.389050] env[63202]: DEBUG oslo_vmware.api [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385960, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.609305] env[63202]: DEBUG nova.compute.manager [req-c8f90298-31f3-4835-b729-82c39a92a050 req-dbee7dc1-a810-40b8-81f2-7bd7c0b8f58d service nova] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Received event network-changed-0193eadb-322b-4278-b259-d56b23555561 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1127.609519] env[63202]: DEBUG nova.compute.manager [req-c8f90298-31f3-4835-b729-82c39a92a050 req-dbee7dc1-a810-40b8-81f2-7bd7c0b8f58d service nova] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Refreshing instance network info cache due to event network-changed-0193eadb-322b-4278-b259-d56b23555561. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1127.609718] env[63202]: DEBUG oslo_concurrency.lockutils [req-c8f90298-31f3-4835-b729-82c39a92a050 req-dbee7dc1-a810-40b8-81f2-7bd7c0b8f58d service nova] Acquiring lock "refresh_cache-ed60fb5e-b08f-4846-942d-c4edde39d1ae" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1127.660305] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Releasing lock "refresh_cache-ed60fb5e-b08f-4846-942d-c4edde39d1ae" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1127.660633] env[63202]: DEBUG nova.compute.manager [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Instance network_info: |[{"id": "0193eadb-322b-4278-b259-d56b23555561", "address": "fa:16:3e:1e:87:93", "network": {"id": "06ab5813-9ad9-4021-9bdb-f2f02af8d73f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1714653503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b2de35030a9484094e964ffc30a822d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3093647a-bab7-4562-ada0-428725e8c0fc", "external-id": "nsx-vlan-transportzone-660", "segmentation_id": 660, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0193eadb-32", "ovs_interfaceid": "0193eadb-322b-4278-b259-d56b23555561", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1127.660949] env[63202]: DEBUG oslo_concurrency.lockutils [req-c8f90298-31f3-4835-b729-82c39a92a050 req-dbee7dc1-a810-40b8-81f2-7bd7c0b8f58d service nova] Acquired lock "refresh_cache-ed60fb5e-b08f-4846-942d-c4edde39d1ae" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1127.661149] env[63202]: DEBUG nova.network.neutron [req-c8f90298-31f3-4835-b729-82c39a92a050 req-dbee7dc1-a810-40b8-81f2-7bd7c0b8f58d service nova] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Refreshing network info cache for port 0193eadb-322b-4278-b259-d56b23555561 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1127.666019] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1e:87:93', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3093647a-bab7-4562-ada0-428725e8c0fc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0193eadb-322b-4278-b259-d56b23555561', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1127.670626] env[63202]: DEBUG oslo.service.loopingcall [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1127.671601] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1127.671844] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ca6294e4-610c-4eaa-962f-934c3a0666a1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.695594] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1127.695594] env[63202]: value = "task-1385961" [ 1127.695594] env[63202]: _type = "Task" [ 1127.695594] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.705509] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385961, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.813737] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 3c6d286d-c152-46f1-b212-96a67324a56d] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1127.892582] env[63202]: DEBUG oslo_vmware.api [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385960, 'name': RemoveSnapshot_Task, 'duration_secs': 0.333381} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.892875] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Deleted Snapshot of the VM instance {{(pid=63202) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1127.893207] env[63202]: DEBUG nova.compute.manager [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1127.894133] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cb35ecc-7398-42f1-881d-7c38d128d9ee {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.205166] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385961, 'name': CreateVM_Task, 'duration_secs': 0.361712} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.207271] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1128.207916] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1128.208361] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1128.208452] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1128.208939] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9c7aeb56-1c93-467c-836c-da2b631f7c76 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.213523] env[63202]: DEBUG oslo_vmware.api [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1128.213523] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52da9bdc-da90-69d7-0244-11db04a0353d" [ 1128.213523] env[63202]: _type = "Task" [ 1128.213523] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.223700] env[63202]: DEBUG oslo_vmware.api [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52da9bdc-da90-69d7-0244-11db04a0353d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.224839] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-250db1ba-9f5e-4459-9667-6d7651d6c7a0 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Volume attach. Driver type: vmdk {{(pid=63202) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1128.225207] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-250db1ba-9f5e-4459-9667-6d7651d6c7a0 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294231', 'volume_id': '71cb14b0-2665-4117-88e9-180e83bde620', 'name': 'volume-71cb14b0-2665-4117-88e9-180e83bde620', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '214c63b6-35e3-46e1-a9de-f3cea251c3c9', 'attached_at': '', 'detached_at': '', 'volume_id': '71cb14b0-2665-4117-88e9-180e83bde620', 'serial': '71cb14b0-2665-4117-88e9-180e83bde620'} {{(pid=63202) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1128.226125] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5b27470-4af9-4e6c-a75b-744558f83177 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.243790] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f0df934-d26a-49e1-bbed-203dda1000c1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.273798] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-250db1ba-9f5e-4459-9667-6d7651d6c7a0 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Reconfiguring VM instance instance-00000067 to attach disk [datastore1] volume-71cb14b0-2665-4117-88e9-180e83bde620/volume-71cb14b0-2665-4117-88e9-180e83bde620.vmdk or device None with type thin {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1128.276712] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-01bb16b1-d1d2-458d-8b0a-3ec8203f7e34 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.298226] env[63202]: DEBUG oslo_vmware.api [None req-250db1ba-9f5e-4459-9667-6d7651d6c7a0 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 1128.298226] env[63202]: value = "task-1385962" [ 1128.298226] env[63202]: _type = "Task" [ 1128.298226] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.307906] env[63202]: DEBUG oslo_vmware.api [None req-250db1ba-9f5e-4459-9667-6d7651d6c7a0 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385962, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.316563] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 0d7fd297-37db-4876-bb91-5bb0b7dc335e] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1128.405975] env[63202]: INFO nova.compute.manager [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Shelve offloading [ 1128.407679] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1128.407952] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cb2ab43c-8c0b-4e4c-bd5d-a1b4a7dd1d78 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.416152] env[63202]: DEBUG oslo_vmware.api [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Waiting for the task: (returnval){ [ 1128.416152] env[63202]: value = "task-1385963" [ 1128.416152] env[63202]: _type = "Task" [ 1128.416152] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.424437] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] VM already powered off {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1128.424621] env[63202]: DEBUG nova.compute.manager [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1128.425385] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ea3b8f8-961a-48ed-b4cb-98a9ed01457d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.430965] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Acquiring lock "refresh_cache-9646810e-06d2-4de0-8f0c-e8719541da53" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1128.431148] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Acquired lock "refresh_cache-9646810e-06d2-4de0-8f0c-e8719541da53" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1128.431342] env[63202]: DEBUG nova.network.neutron [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1128.436297] env[63202]: DEBUG nova.network.neutron [req-c8f90298-31f3-4835-b729-82c39a92a050 req-dbee7dc1-a810-40b8-81f2-7bd7c0b8f58d service nova] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Updated VIF entry in instance network info cache for port 0193eadb-322b-4278-b259-d56b23555561. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1128.436686] env[63202]: DEBUG nova.network.neutron [req-c8f90298-31f3-4835-b729-82c39a92a050 req-dbee7dc1-a810-40b8-81f2-7bd7c0b8f58d service nova] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Updating instance_info_cache with network_info: [{"id": "0193eadb-322b-4278-b259-d56b23555561", "address": "fa:16:3e:1e:87:93", "network": {"id": "06ab5813-9ad9-4021-9bdb-f2f02af8d73f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1714653503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b2de35030a9484094e964ffc30a822d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3093647a-bab7-4562-ada0-428725e8c0fc", "external-id": "nsx-vlan-transportzone-660", "segmentation_id": 660, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0193eadb-32", "ovs_interfaceid": "0193eadb-322b-4278-b259-d56b23555561", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1128.723607] env[63202]: DEBUG oslo_vmware.api [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52da9bdc-da90-69d7-0244-11db04a0353d, 'name': SearchDatastore_Task, 'duration_secs': 0.016549} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.723945] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1128.724187] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1128.724399] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1128.724553] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1128.724730] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1128.724974] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-87798b19-5699-43cd-a806-f01e5a076387 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.744085] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1128.744271] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1128.744956] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-875e0933-4211-407e-b928-f6bc5eb811b3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.749476] env[63202]: DEBUG oslo_vmware.api [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1128.749476] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]522918b2-6d0c-4dad-b677-00db67a0a6fb" [ 1128.749476] env[63202]: _type = "Task" [ 1128.749476] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.756538] env[63202]: DEBUG oslo_vmware.api [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]522918b2-6d0c-4dad-b677-00db67a0a6fb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.807279] env[63202]: DEBUG oslo_vmware.api [None req-250db1ba-9f5e-4459-9667-6d7651d6c7a0 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385962, 'name': ReconfigVM_Task, 'duration_secs': 0.35338} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.807512] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-250db1ba-9f5e-4459-9667-6d7651d6c7a0 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Reconfigured VM instance instance-00000067 to attach disk [datastore1] volume-71cb14b0-2665-4117-88e9-180e83bde620/volume-71cb14b0-2665-4117-88e9-180e83bde620.vmdk or device None with type thin {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1128.812286] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-27c6a5c1-b16a-4016-aa96-9f77e245b9ca {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.821743] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: f23f2c04-c266-46a2-ad3f-34100f6246a0] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1128.828521] env[63202]: DEBUG oslo_vmware.api [None req-250db1ba-9f5e-4459-9667-6d7651d6c7a0 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 1128.828521] env[63202]: value = "task-1385964" [ 1128.828521] env[63202]: _type = "Task" [ 1128.828521] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.835886] env[63202]: DEBUG oslo_vmware.api [None req-250db1ba-9f5e-4459-9667-6d7651d6c7a0 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385964, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.939440] env[63202]: DEBUG oslo_concurrency.lockutils [req-c8f90298-31f3-4835-b729-82c39a92a050 req-dbee7dc1-a810-40b8-81f2-7bd7c0b8f58d service nova] Releasing lock "refresh_cache-ed60fb5e-b08f-4846-942d-c4edde39d1ae" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1129.119940] env[63202]: DEBUG nova.network.neutron [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Updating instance_info_cache with network_info: [{"id": "55537461-97ca-40a4-8086-69473669d5b6", "address": "fa:16:3e:cc:19:9e", "network": {"id": "f105b723-b2a9-401d-a936-e31b4b6609f6", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-298435423-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "433b1d11832147b886bed7a3a4952768", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e4c8c8fd-baca-4e60-97dc-ff0418d63215", "external-id": "nsx-vlan-transportzone-178", "segmentation_id": 178, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap55537461-97", "ovs_interfaceid": "55537461-97ca-40a4-8086-69473669d5b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1129.260175] env[63202]: DEBUG oslo_vmware.api [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]522918b2-6d0c-4dad-b677-00db67a0a6fb, 'name': SearchDatastore_Task, 'duration_secs': 0.008237} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.260901] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3cd17afc-1c1b-421d-b13d-daddac0ce44a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.265893] env[63202]: DEBUG oslo_vmware.api [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1129.265893] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52dd63ff-2ff1-e786-7464-1e84af7624df" [ 1129.265893] env[63202]: _type = "Task" [ 1129.265893] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.275345] env[63202]: DEBUG oslo_vmware.api [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52dd63ff-2ff1-e786-7464-1e84af7624df, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.325012] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 8b5b30b9-f1d9-4bb9-ab1e-e5ea689666e8] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1129.337304] env[63202]: DEBUG oslo_vmware.api [None req-250db1ba-9f5e-4459-9667-6d7651d6c7a0 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385964, 'name': ReconfigVM_Task, 'duration_secs': 0.153445} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.337612] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-250db1ba-9f5e-4459-9667-6d7651d6c7a0 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294231', 'volume_id': '71cb14b0-2665-4117-88e9-180e83bde620', 'name': 'volume-71cb14b0-2665-4117-88e9-180e83bde620', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '214c63b6-35e3-46e1-a9de-f3cea251c3c9', 'attached_at': '', 'detached_at': '', 'volume_id': '71cb14b0-2665-4117-88e9-180e83bde620', 'serial': '71cb14b0-2665-4117-88e9-180e83bde620'} {{(pid=63202) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1129.622453] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Releasing lock "refresh_cache-9646810e-06d2-4de0-8f0c-e8719541da53" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1129.776757] env[63202]: DEBUG oslo_vmware.api [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52dd63ff-2ff1-e786-7464-1e84af7624df, 'name': SearchDatastore_Task, 'duration_secs': 0.009514} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.777038] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1129.777300] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] ed60fb5e-b08f-4846-942d-c4edde39d1ae/ed60fb5e-b08f-4846-942d-c4edde39d1ae.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1129.777553] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0374ee8a-b6e6-4226-951c-b066da35ad73 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.784669] env[63202]: DEBUG oslo_vmware.api [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1129.784669] env[63202]: value = "task-1385965" [ 1129.784669] env[63202]: _type = "Task" [ 1129.784669] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.792118] env[63202]: DEBUG oslo_vmware.api [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385965, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.827890] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: dd146c71-c391-41e8-8cc7-7276f4a518d9] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1129.877515] env[63202]: DEBUG nova.compute.manager [req-937e04cd-8a96-4727-991f-a1e162252f4e req-faff21d0-e95b-44c1-b1e5-10ca356e1b5a service nova] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Received event network-vif-unplugged-55537461-97ca-40a4-8086-69473669d5b6 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1129.877659] env[63202]: DEBUG oslo_concurrency.lockutils [req-937e04cd-8a96-4727-991f-a1e162252f4e req-faff21d0-e95b-44c1-b1e5-10ca356e1b5a service nova] Acquiring lock "9646810e-06d2-4de0-8f0c-e8719541da53-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1129.877865] env[63202]: DEBUG oslo_concurrency.lockutils [req-937e04cd-8a96-4727-991f-a1e162252f4e req-faff21d0-e95b-44c1-b1e5-10ca356e1b5a service nova] Lock "9646810e-06d2-4de0-8f0c-e8719541da53-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1129.878050] env[63202]: DEBUG oslo_concurrency.lockutils [req-937e04cd-8a96-4727-991f-a1e162252f4e req-faff21d0-e95b-44c1-b1e5-10ca356e1b5a service nova] Lock "9646810e-06d2-4de0-8f0c-e8719541da53-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1129.878274] env[63202]: DEBUG nova.compute.manager [req-937e04cd-8a96-4727-991f-a1e162252f4e req-faff21d0-e95b-44c1-b1e5-10ca356e1b5a service nova] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] No waiting events found dispatching network-vif-unplugged-55537461-97ca-40a4-8086-69473669d5b6 {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1129.878396] env[63202]: WARNING nova.compute.manager [req-937e04cd-8a96-4727-991f-a1e162252f4e req-faff21d0-e95b-44c1-b1e5-10ca356e1b5a service nova] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Received unexpected event network-vif-unplugged-55537461-97ca-40a4-8086-69473669d5b6 for instance with vm_state shelved and task_state shelving_offloading. [ 1129.900344] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1129.901429] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7f5b540-ed78-429d-b710-f517a397cf72 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.909049] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1129.909202] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bb615a3b-4a35-4cf3-b20f-2038aa983b38 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.979504] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1129.979827] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Deleting contents of the VM from datastore datastore2 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1129.980095] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Deleting the datastore file [datastore2] 9646810e-06d2-4de0-8f0c-e8719541da53 {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1129.980444] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fd86b4da-6fec-40d8-ba6d-bb2b803a69a7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.987745] env[63202]: DEBUG oslo_vmware.api [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Waiting for the task: (returnval){ [ 1129.987745] env[63202]: value = "task-1385967" [ 1129.987745] env[63202]: _type = "Task" [ 1129.987745] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.996431] env[63202]: DEBUG oslo_vmware.api [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385967, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.293797] env[63202]: DEBUG oslo_vmware.api [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385965, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.475469} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.294072] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] ed60fb5e-b08f-4846-942d-c4edde39d1ae/ed60fb5e-b08f-4846-942d-c4edde39d1ae.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1130.294297] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1130.294548] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-04362899-c4d4-4b89-b91b-d4ae96af5ebb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.300467] env[63202]: DEBUG oslo_vmware.api [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1130.300467] env[63202]: value = "task-1385968" [ 1130.300467] env[63202]: _type = "Task" [ 1130.300467] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.307781] env[63202]: DEBUG oslo_vmware.api [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385968, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.331348] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: eca51705-a972-48f6-85f5-6c397dad955d] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1130.376025] env[63202]: DEBUG nova.objects.instance [None req-250db1ba-9f5e-4459-9667-6d7651d6c7a0 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lazy-loading 'flavor' on Instance uuid 214c63b6-35e3-46e1-a9de-f3cea251c3c9 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1130.497206] env[63202]: DEBUG oslo_vmware.api [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385967, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.264835} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.497462] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1130.497649] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Deleted contents of the VM from datastore datastore2 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1130.497822] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1130.647112] env[63202]: INFO nova.scheduler.client.report [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Deleted allocations for instance 9646810e-06d2-4de0-8f0c-e8719541da53 [ 1130.809986] env[63202]: DEBUG oslo_vmware.api [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385968, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.05356} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.810286] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1130.811057] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a2e1bba-5192-426e-b906-b354de2d077e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.833138] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] ed60fb5e-b08f-4846-942d-c4edde39d1ae/ed60fb5e-b08f-4846-942d-c4edde39d1ae.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1130.833457] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-de0daafd-9532-44e3-8f30-66db8174b6b1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.847327] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 8437f856-a707-49c5-b8eb-5a22cdb990f8] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1130.854754] env[63202]: DEBUG oslo_vmware.api [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1130.854754] env[63202]: value = "task-1385969" [ 1130.854754] env[63202]: _type = "Task" [ 1130.854754] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.861913] env[63202]: DEBUG oslo_vmware.api [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385969, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.880271] env[63202]: DEBUG oslo_concurrency.lockutils [None req-250db1ba-9f5e-4459-9667-6d7651d6c7a0 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "214c63b6-35e3-46e1-a9de-f3cea251c3c9" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.266s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1131.095902] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89f7cbd5-5fa6-4737-a135-7e384c69f672 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquiring lock "214c63b6-35e3-46e1-a9de-f3cea251c3c9" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1131.097456] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89f7cbd5-5fa6-4737-a135-7e384c69f672 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "214c63b6-35e3-46e1-a9de-f3cea251c3c9" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1131.151359] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1131.151626] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1131.151846] env[63202]: DEBUG nova.objects.instance [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Lazy-loading 'resources' on Instance uuid 9646810e-06d2-4de0-8f0c-e8719541da53 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1131.350538] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: da285417-bd38-4387-8521-df326ca0e326] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1131.364867] env[63202]: DEBUG oslo_vmware.api [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385969, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.601199] env[63202]: INFO nova.compute.manager [None req-89f7cbd5-5fa6-4737-a135-7e384c69f672 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Detaching volume 71cb14b0-2665-4117-88e9-180e83bde620 [ 1131.639805] env[63202]: INFO nova.virt.block_device [None req-89f7cbd5-5fa6-4737-a135-7e384c69f672 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Attempting to driver detach volume 71cb14b0-2665-4117-88e9-180e83bde620 from mountpoint /dev/sdb [ 1131.640076] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-89f7cbd5-5fa6-4737-a135-7e384c69f672 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Volume detach. Driver type: vmdk {{(pid=63202) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1131.640308] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-89f7cbd5-5fa6-4737-a135-7e384c69f672 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294231', 'volume_id': '71cb14b0-2665-4117-88e9-180e83bde620', 'name': 'volume-71cb14b0-2665-4117-88e9-180e83bde620', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '214c63b6-35e3-46e1-a9de-f3cea251c3c9', 'attached_at': '', 'detached_at': '', 'volume_id': '71cb14b0-2665-4117-88e9-180e83bde620', 'serial': '71cb14b0-2665-4117-88e9-180e83bde620'} {{(pid=63202) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1131.641289] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe4a7d29-e071-44fa-90e6-1acd07a0c9cd {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.674342] env[63202]: DEBUG nova.objects.instance [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Lazy-loading 'numa_topology' on Instance uuid 9646810e-06d2-4de0-8f0c-e8719541da53 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1131.676585] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96767b31-a57d-462c-8284-358071e3e349 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.698083] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d4acd1d-3ad0-42e2-a195-fbc7e918e942 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.721859] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f03afa11-4599-4b23-b6fc-5dc36f313fea {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.741687] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-89f7cbd5-5fa6-4737-a135-7e384c69f672 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] The volume has not been displaced from its original location: [datastore1] volume-71cb14b0-2665-4117-88e9-180e83bde620/volume-71cb14b0-2665-4117-88e9-180e83bde620.vmdk. No consolidation needed. {{(pid=63202) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1131.746886] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-89f7cbd5-5fa6-4737-a135-7e384c69f672 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Reconfiguring VM instance instance-00000067 to detach disk 2001 {{(pid=63202) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1131.747402] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ffc59f4f-fe22-4011-b8d8-546eede97154 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.764722] env[63202]: DEBUG oslo_vmware.api [None req-89f7cbd5-5fa6-4737-a135-7e384c69f672 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 1131.764722] env[63202]: value = "task-1385970" [ 1131.764722] env[63202]: _type = "Task" [ 1131.764722] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.771895] env[63202]: DEBUG oslo_vmware.api [None req-89f7cbd5-5fa6-4737-a135-7e384c69f672 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385970, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.853624] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 2110dfa7-1795-4c8f-92bf-18921409c99f] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1131.863664] env[63202]: DEBUG oslo_vmware.api [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385969, 'name': ReconfigVM_Task, 'duration_secs': 0.549142} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.863916] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Reconfigured VM instance instance-0000006a to attach disk [datastore1] ed60fb5e-b08f-4846-942d-c4edde39d1ae/ed60fb5e-b08f-4846-942d-c4edde39d1ae.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1131.864523] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-778edf5c-571d-42d8-bacf-170149d99295 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.870232] env[63202]: DEBUG oslo_vmware.api [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1131.870232] env[63202]: value = "task-1385971" [ 1131.870232] env[63202]: _type = "Task" [ 1131.870232] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.879287] env[63202]: DEBUG oslo_vmware.api [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385971, 'name': Rename_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.905419] env[63202]: DEBUG nova.compute.manager [req-ea89c891-cd17-4451-a8fa-112dcf6fbcf4 req-35eaf240-6326-4979-a8c5-3dfa6ffdb5ad service nova] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Received event network-changed-55537461-97ca-40a4-8086-69473669d5b6 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1131.905799] env[63202]: DEBUG nova.compute.manager [req-ea89c891-cd17-4451-a8fa-112dcf6fbcf4 req-35eaf240-6326-4979-a8c5-3dfa6ffdb5ad service nova] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Refreshing instance network info cache due to event network-changed-55537461-97ca-40a4-8086-69473669d5b6. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1131.905950] env[63202]: DEBUG oslo_concurrency.lockutils [req-ea89c891-cd17-4451-a8fa-112dcf6fbcf4 req-35eaf240-6326-4979-a8c5-3dfa6ffdb5ad service nova] Acquiring lock "refresh_cache-9646810e-06d2-4de0-8f0c-e8719541da53" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1131.906102] env[63202]: DEBUG oslo_concurrency.lockutils [req-ea89c891-cd17-4451-a8fa-112dcf6fbcf4 req-35eaf240-6326-4979-a8c5-3dfa6ffdb5ad service nova] Acquired lock "refresh_cache-9646810e-06d2-4de0-8f0c-e8719541da53" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.906275] env[63202]: DEBUG nova.network.neutron [req-ea89c891-cd17-4451-a8fa-112dcf6fbcf4 req-35eaf240-6326-4979-a8c5-3dfa6ffdb5ad service nova] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Refreshing network info cache for port 55537461-97ca-40a4-8086-69473669d5b6 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1132.184858] env[63202]: DEBUG nova.objects.base [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Object Instance<9646810e-06d2-4de0-8f0c-e8719541da53> lazy-loaded attributes: resources,numa_topology {{(pid=63202) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1132.229559] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Acquiring lock "9646810e-06d2-4de0-8f0c-e8719541da53" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.238305] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b56b917d-1dc1-4084-bd73-3f631184489e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.245999] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a07fef6f-2380-4c2e-bb85-51d40e744d0a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.280266] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bab48c1-ab54-4551-b235-6f49821d0b32 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.287179] env[63202]: DEBUG oslo_vmware.api [None req-89f7cbd5-5fa6-4737-a135-7e384c69f672 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385970, 'name': ReconfigVM_Task, 'duration_secs': 0.203065} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.289219] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-89f7cbd5-5fa6-4737-a135-7e384c69f672 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Reconfigured VM instance instance-00000067 to detach disk 2001 {{(pid=63202) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1132.293814] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eeb4fcfa-f1c0-4b09-a6a0-f3a123988782 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.304039] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2854f28-59c4-42f6-baa8-e3d07a239e6a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.316980] env[63202]: DEBUG nova.compute.provider_tree [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1132.319200] env[63202]: DEBUG oslo_vmware.api [None req-89f7cbd5-5fa6-4737-a135-7e384c69f672 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 1132.319200] env[63202]: value = "task-1385972" [ 1132.319200] env[63202]: _type = "Task" [ 1132.319200] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.326415] env[63202]: DEBUG oslo_vmware.api [None req-89f7cbd5-5fa6-4737-a135-7e384c69f672 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385972, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.359709] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: a961b2a0-39c1-4267-9229-068e2b6ecd67] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1132.379757] env[63202]: DEBUG oslo_vmware.api [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385971, 'name': Rename_Task, 'duration_secs': 0.130464} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.380053] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1132.380321] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5fc443e4-28bc-4bca-8095-79e9d01b9d95 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.386948] env[63202]: DEBUG oslo_vmware.api [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1132.386948] env[63202]: value = "task-1385973" [ 1132.386948] env[63202]: _type = "Task" [ 1132.386948] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.395813] env[63202]: DEBUG oslo_vmware.api [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385973, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.614029] env[63202]: DEBUG nova.network.neutron [req-ea89c891-cd17-4451-a8fa-112dcf6fbcf4 req-35eaf240-6326-4979-a8c5-3dfa6ffdb5ad service nova] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Updated VIF entry in instance network info cache for port 55537461-97ca-40a4-8086-69473669d5b6. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1132.614644] env[63202]: DEBUG nova.network.neutron [req-ea89c891-cd17-4451-a8fa-112dcf6fbcf4 req-35eaf240-6326-4979-a8c5-3dfa6ffdb5ad service nova] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Updating instance_info_cache with network_info: [{"id": "55537461-97ca-40a4-8086-69473669d5b6", "address": "fa:16:3e:cc:19:9e", "network": {"id": "f105b723-b2a9-401d-a936-e31b4b6609f6", "bridge": null, "label": "tempest-ServersNegativeTestJSON-298435423-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "433b1d11832147b886bed7a3a4952768", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap55537461-97", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1132.821420] env[63202]: DEBUG nova.scheduler.client.report [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1132.833204] env[63202]: DEBUG oslo_vmware.api [None req-89f7cbd5-5fa6-4737-a135-7e384c69f672 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385972, 'name': ReconfigVM_Task, 'duration_secs': 0.173992} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.833498] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-89f7cbd5-5fa6-4737-a135-7e384c69f672 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294231', 'volume_id': '71cb14b0-2665-4117-88e9-180e83bde620', 'name': 'volume-71cb14b0-2665-4117-88e9-180e83bde620', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '214c63b6-35e3-46e1-a9de-f3cea251c3c9', 'attached_at': '', 'detached_at': '', 'volume_id': '71cb14b0-2665-4117-88e9-180e83bde620', 'serial': '71cb14b0-2665-4117-88e9-180e83bde620'} {{(pid=63202) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1132.862863] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 7437595c-fa35-483e-95f3-b75405b6bd13] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1132.896735] env[63202]: DEBUG oslo_vmware.api [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385973, 'name': PowerOnVM_Task, 'duration_secs': 0.447114} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.896997] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1132.897228] env[63202]: INFO nova.compute.manager [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Took 7.87 seconds to spawn the instance on the hypervisor. [ 1132.897410] env[63202]: DEBUG nova.compute.manager [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1132.898191] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f008a5e9-e7ec-4772-90d4-433052f826cb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.117965] env[63202]: DEBUG oslo_concurrency.lockutils [req-ea89c891-cd17-4451-a8fa-112dcf6fbcf4 req-35eaf240-6326-4979-a8c5-3dfa6ffdb5ad service nova] Releasing lock "refresh_cache-9646810e-06d2-4de0-8f0c-e8719541da53" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1133.329916] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.178s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1133.366369] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: f3128c09-3680-4b0b-b463-3d6cd203fcf4] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1133.374197] env[63202]: DEBUG nova.objects.instance [None req-89f7cbd5-5fa6-4737-a135-7e384c69f672 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lazy-loading 'flavor' on Instance uuid 214c63b6-35e3-46e1-a9de-f3cea251c3c9 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1133.413228] env[63202]: INFO nova.compute.manager [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Took 12.57 seconds to build instance. [ 1133.843491] env[63202]: DEBUG oslo_concurrency.lockutils [None req-e6224835-fbe4-4bcd-b13d-fa60980b688f tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Lock "9646810e-06d2-4de0-8f0c-e8719541da53" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 19.729s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1133.845427] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Lock "9646810e-06d2-4de0-8f0c-e8719541da53" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.615s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1133.845427] env[63202]: INFO nova.compute.manager [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Unshelving [ 1133.869025] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 8755bceb-d510-4429-bd98-d6a63faf739d] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1133.914802] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5f2db8d2-bf88-4936-93cc-174ec5dea46a tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lock "ed60fb5e-b08f-4846-942d-c4edde39d1ae" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.079s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1134.047256] env[63202]: DEBUG nova.compute.manager [req-68833fbc-9548-4abe-9c2a-e1b2645b5228 req-aefd2e94-2c23-47ab-af4a-eec4b22e6ee1 service nova] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Received event network-changed-0193eadb-322b-4278-b259-d56b23555561 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1134.047835] env[63202]: DEBUG nova.compute.manager [req-68833fbc-9548-4abe-9c2a-e1b2645b5228 req-aefd2e94-2c23-47ab-af4a-eec4b22e6ee1 service nova] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Refreshing instance network info cache due to event network-changed-0193eadb-322b-4278-b259-d56b23555561. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1134.047835] env[63202]: DEBUG oslo_concurrency.lockutils [req-68833fbc-9548-4abe-9c2a-e1b2645b5228 req-aefd2e94-2c23-47ab-af4a-eec4b22e6ee1 service nova] Acquiring lock "refresh_cache-ed60fb5e-b08f-4846-942d-c4edde39d1ae" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1134.047835] env[63202]: DEBUG oslo_concurrency.lockutils [req-68833fbc-9548-4abe-9c2a-e1b2645b5228 req-aefd2e94-2c23-47ab-af4a-eec4b22e6ee1 service nova] Acquired lock "refresh_cache-ed60fb5e-b08f-4846-942d-c4edde39d1ae" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1134.048197] env[63202]: DEBUG nova.network.neutron [req-68833fbc-9548-4abe-9c2a-e1b2645b5228 req-aefd2e94-2c23-47ab-af4a-eec4b22e6ee1 service nova] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Refreshing network info cache for port 0193eadb-322b-4278-b259-d56b23555561 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1134.371475] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: d0d6e380-9337-4f69-8434-6a33ac8a33d6] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1134.383993] env[63202]: DEBUG oslo_concurrency.lockutils [None req-89f7cbd5-5fa6-4737-a135-7e384c69f672 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "214c63b6-35e3-46e1-a9de-f3cea251c3c9" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.288s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1134.778152] env[63202]: DEBUG nova.network.neutron [req-68833fbc-9548-4abe-9c2a-e1b2645b5228 req-aefd2e94-2c23-47ab-af4a-eec4b22e6ee1 service nova] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Updated VIF entry in instance network info cache for port 0193eadb-322b-4278-b259-d56b23555561. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1134.778600] env[63202]: DEBUG nova.network.neutron [req-68833fbc-9548-4abe-9c2a-e1b2645b5228 req-aefd2e94-2c23-47ab-af4a-eec4b22e6ee1 service nova] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Updating instance_info_cache with network_info: [{"id": "0193eadb-322b-4278-b259-d56b23555561", "address": "fa:16:3e:1e:87:93", "network": {"id": "06ab5813-9ad9-4021-9bdb-f2f02af8d73f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1714653503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.182", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b2de35030a9484094e964ffc30a822d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3093647a-bab7-4562-ada0-428725e8c0fc", "external-id": "nsx-vlan-transportzone-660", "segmentation_id": 660, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0193eadb-32", "ovs_interfaceid": "0193eadb-322b-4278-b259-d56b23555561", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1134.873823] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1134.874264] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1134.874629] env[63202]: DEBUG nova.objects.instance [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Lazy-loading 'pci_requests' on Instance uuid 9646810e-06d2-4de0-8f0c-e8719541da53 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1134.876157] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: fc7d3924-a624-4ea4-890c-7628595fb733] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1135.281937] env[63202]: DEBUG oslo_concurrency.lockutils [req-68833fbc-9548-4abe-9c2a-e1b2645b5228 req-aefd2e94-2c23-47ab-af4a-eec4b22e6ee1 service nova] Releasing lock "refresh_cache-ed60fb5e-b08f-4846-942d-c4edde39d1ae" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1135.350918] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1c60580d-b2ad-40ca-bb8b-827c4458e4c9 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquiring lock "214c63b6-35e3-46e1-a9de-f3cea251c3c9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1135.351227] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1c60580d-b2ad-40ca-bb8b-827c4458e4c9 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "214c63b6-35e3-46e1-a9de-f3cea251c3c9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1135.351448] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1c60580d-b2ad-40ca-bb8b-827c4458e4c9 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquiring lock "214c63b6-35e3-46e1-a9de-f3cea251c3c9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1135.351644] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1c60580d-b2ad-40ca-bb8b-827c4458e4c9 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "214c63b6-35e3-46e1-a9de-f3cea251c3c9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1135.351835] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1c60580d-b2ad-40ca-bb8b-827c4458e4c9 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "214c63b6-35e3-46e1-a9de-f3cea251c3c9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1135.354191] env[63202]: INFO nova.compute.manager [None req-1c60580d-b2ad-40ca-bb8b-827c4458e4c9 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Terminating instance [ 1135.356026] env[63202]: DEBUG nova.compute.manager [None req-1c60580d-b2ad-40ca-bb8b-827c4458e4c9 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1135.356223] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-1c60580d-b2ad-40ca-bb8b-827c4458e4c9 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1135.357098] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89c7c98d-743f-489c-b11c-940dd6f6b685 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.365583] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c60580d-b2ad-40ca-bb8b-827c4458e4c9 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1135.365857] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-178ee92c-cdfb-4b90-9975-c55f513e213f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.372273] env[63202]: DEBUG oslo_vmware.api [None req-1c60580d-b2ad-40ca-bb8b-827c4458e4c9 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 1135.372273] env[63202]: value = "task-1385974" [ 1135.372273] env[63202]: _type = "Task" [ 1135.372273] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.383021] env[63202]: DEBUG nova.objects.instance [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Lazy-loading 'numa_topology' on Instance uuid 9646810e-06d2-4de0-8f0c-e8719541da53 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1135.384313] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: f3652744-e072-4700-80d4-b9eca414c5cb] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1135.386236] env[63202]: DEBUG oslo_vmware.api [None req-1c60580d-b2ad-40ca-bb8b-827c4458e4c9 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385974, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.886749] env[63202]: DEBUG oslo_vmware.api [None req-1c60580d-b2ad-40ca-bb8b-827c4458e4c9 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385974, 'name': PowerOffVM_Task, 'duration_secs': 0.207405} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.887256] env[63202]: INFO nova.compute.claims [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1135.889766] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: b1bccea9-2d79-431a-8be0-0a5ab293542a] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1135.891320] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c60580d-b2ad-40ca-bb8b-827c4458e4c9 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1135.891499] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-1c60580d-b2ad-40ca-bb8b-827c4458e4c9 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1135.892056] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4a8aaa2f-6d34-45d9-bb5e-00596b22f0fb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.080460] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-1c60580d-b2ad-40ca-bb8b-827c4458e4c9 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1136.080730] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-1c60580d-b2ad-40ca-bb8b-827c4458e4c9 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Deleting contents of the VM from datastore datastore2 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1136.080917] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c60580d-b2ad-40ca-bb8b-827c4458e4c9 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Deleting the datastore file [datastore2] 214c63b6-35e3-46e1-a9de-f3cea251c3c9 {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1136.081211] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1a042254-0263-4435-a74d-9f1d200ee8b7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.088101] env[63202]: DEBUG oslo_vmware.api [None req-1c60580d-b2ad-40ca-bb8b-827c4458e4c9 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 1136.088101] env[63202]: value = "task-1385976" [ 1136.088101] env[63202]: _type = "Task" [ 1136.088101] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.096410] env[63202]: DEBUG oslo_vmware.api [None req-1c60580d-b2ad-40ca-bb8b-827c4458e4c9 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385976, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.395509] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: af2bedc2-28ee-4679-ae38-1cceb2af05d6] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1136.598166] env[63202]: DEBUG oslo_vmware.api [None req-1c60580d-b2ad-40ca-bb8b-827c4458e4c9 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385976, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.210287} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.598411] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c60580d-b2ad-40ca-bb8b-827c4458e4c9 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1136.598596] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-1c60580d-b2ad-40ca-bb8b-827c4458e4c9 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Deleted contents of the VM from datastore datastore2 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1136.598774] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-1c60580d-b2ad-40ca-bb8b-827c4458e4c9 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1136.598948] env[63202]: INFO nova.compute.manager [None req-1c60580d-b2ad-40ca-bb8b-827c4458e4c9 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Took 1.24 seconds to destroy the instance on the hypervisor. [ 1136.599210] env[63202]: DEBUG oslo.service.loopingcall [None req-1c60580d-b2ad-40ca-bb8b-827c4458e4c9 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1136.599406] env[63202]: DEBUG nova.compute.manager [-] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1136.599503] env[63202]: DEBUG nova.network.neutron [-] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1136.902512] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 753d190b-f4a4-4438-bc98-94564e3ec73d] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1136.958355] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dc0f35b-a2f9-45b1-b0a0-e1c4c374db15 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.966861] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49cb4c22-6330-44e3-a8d1-ecc5a77eb699 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.997246] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bb37de4-27da-46a7-8858-8720477c6e2b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.004810] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c974bd98-dd15-4178-be3d-8551bd755fc5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.018116] env[63202]: DEBUG nova.compute.provider_tree [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1137.042407] env[63202]: DEBUG nova.compute.manager [req-b56c6d5b-e2ff-466c-bf3d-f8a7004addcc req-32f5cc34-7513-41d6-ac0a-d0a6a9cac30a service nova] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Received event network-vif-deleted-76e223aa-1b1a-4d12-a8a6-89bb986656b5 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1137.042637] env[63202]: INFO nova.compute.manager [req-b56c6d5b-e2ff-466c-bf3d-f8a7004addcc req-32f5cc34-7513-41d6-ac0a-d0a6a9cac30a service nova] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Neutron deleted interface 76e223aa-1b1a-4d12-a8a6-89bb986656b5; detaching it from the instance and deleting it from the info cache [ 1137.042855] env[63202]: DEBUG nova.network.neutron [req-b56c6d5b-e2ff-466c-bf3d-f8a7004addcc req-32f5cc34-7513-41d6-ac0a-d0a6a9cac30a service nova] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1137.407468] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: e775e5e1-521a-4fc7-80e6-bcb6a70516c5] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1137.522317] env[63202]: DEBUG nova.scheduler.client.report [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1137.525405] env[63202]: DEBUG nova.network.neutron [-] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1137.545893] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1ba55ffe-39c1-40b7-ad93-905b65459da6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.556266] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ef7c2e1-0f86-4ec6-9963-47c69e01752f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.579596] env[63202]: DEBUG nova.compute.manager [req-b56c6d5b-e2ff-466c-bf3d-f8a7004addcc req-32f5cc34-7513-41d6-ac0a-d0a6a9cac30a service nova] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Detach interface failed, port_id=76e223aa-1b1a-4d12-a8a6-89bb986656b5, reason: Instance 214c63b6-35e3-46e1-a9de-f3cea251c3c9 could not be found. {{(pid=63202) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1137.910821] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 7a72a1ab-3b3d-450d-a3a6-572b0035eb0e] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1138.028949] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.155s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.031142] env[63202]: INFO nova.compute.manager [-] [instance: 214c63b6-35e3-46e1-a9de-f3cea251c3c9] Took 1.43 seconds to deallocate network for instance. [ 1138.059838] env[63202]: INFO nova.network.neutron [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Updating port 55537461-97ca-40a4-8086-69473669d5b6 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1138.414086] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 61a6528e-4e4e-49b7-be7b-dd30bd8023d8] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1138.537865] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1c60580d-b2ad-40ca-bb8b-827c4458e4c9 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.538114] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1c60580d-b2ad-40ca-bb8b-827c4458e4c9 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1138.538344] env[63202]: DEBUG nova.objects.instance [None req-1c60580d-b2ad-40ca-bb8b-827c4458e4c9 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lazy-loading 'resources' on Instance uuid 214c63b6-35e3-46e1-a9de-f3cea251c3c9 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1138.916977] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 3d1518d3-bed2-4b7b-af1d-d7da49e92874] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1139.089835] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e1a4ded-fb79-4bee-997e-74870ca72b3d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.097753] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7057ceb7-6309-4aca-9302-2c3e2dc86051 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.126541] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9b1f60e-740c-4b98-81d9-7194b809b330 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.133259] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97a63c07-6db7-47be-b87f-a5c0d6f3ef5f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.145608] env[63202]: DEBUG nova.compute.provider_tree [None req-1c60580d-b2ad-40ca-bb8b-827c4458e4c9 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1139.421591] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 36291628-af7d-43cf-8149-09f57df47890] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1139.427903] env[63202]: DEBUG nova.compute.manager [req-da7a1788-a451-494a-9557-6db8a88c3a2f req-38ec19f5-26bb-4dc5-8399-9954015be8bc service nova] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Received event network-vif-plugged-55537461-97ca-40a4-8086-69473669d5b6 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1139.427903] env[63202]: DEBUG oslo_concurrency.lockutils [req-da7a1788-a451-494a-9557-6db8a88c3a2f req-38ec19f5-26bb-4dc5-8399-9954015be8bc service nova] Acquiring lock "9646810e-06d2-4de0-8f0c-e8719541da53-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1139.427903] env[63202]: DEBUG oslo_concurrency.lockutils [req-da7a1788-a451-494a-9557-6db8a88c3a2f req-38ec19f5-26bb-4dc5-8399-9954015be8bc service nova] Lock "9646810e-06d2-4de0-8f0c-e8719541da53-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1139.427903] env[63202]: DEBUG oslo_concurrency.lockutils [req-da7a1788-a451-494a-9557-6db8a88c3a2f req-38ec19f5-26bb-4dc5-8399-9954015be8bc service nova] Lock "9646810e-06d2-4de0-8f0c-e8719541da53-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1139.427903] env[63202]: DEBUG nova.compute.manager [req-da7a1788-a451-494a-9557-6db8a88c3a2f req-38ec19f5-26bb-4dc5-8399-9954015be8bc service nova] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] No waiting events found dispatching network-vif-plugged-55537461-97ca-40a4-8086-69473669d5b6 {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1139.428272] env[63202]: WARNING nova.compute.manager [req-da7a1788-a451-494a-9557-6db8a88c3a2f req-38ec19f5-26bb-4dc5-8399-9954015be8bc service nova] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Received unexpected event network-vif-plugged-55537461-97ca-40a4-8086-69473669d5b6 for instance with vm_state shelved_offloaded and task_state spawning. [ 1139.514304] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Acquiring lock "refresh_cache-9646810e-06d2-4de0-8f0c-e8719541da53" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1139.514798] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Acquired lock "refresh_cache-9646810e-06d2-4de0-8f0c-e8719541da53" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1139.515679] env[63202]: DEBUG nova.network.neutron [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1139.648642] env[63202]: DEBUG nova.scheduler.client.report [None req-1c60580d-b2ad-40ca-bb8b-827c4458e4c9 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1139.925517] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 2243bb69-0dc5-49cd-b94e-73e703cbadc0] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1140.153105] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1c60580d-b2ad-40ca-bb8b-827c4458e4c9 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.615s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1140.170269] env[63202]: INFO nova.scheduler.client.report [None req-1c60580d-b2ad-40ca-bb8b-827c4458e4c9 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Deleted allocations for instance 214c63b6-35e3-46e1-a9de-f3cea251c3c9 [ 1140.204031] env[63202]: DEBUG nova.network.neutron [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Updating instance_info_cache with network_info: [{"id": "55537461-97ca-40a4-8086-69473669d5b6", "address": "fa:16:3e:cc:19:9e", "network": {"id": "f105b723-b2a9-401d-a936-e31b4b6609f6", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-298435423-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "433b1d11832147b886bed7a3a4952768", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e4c8c8fd-baca-4e60-97dc-ff0418d63215", "external-id": "nsx-vlan-transportzone-178", "segmentation_id": 178, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap55537461-97", "ovs_interfaceid": "55537461-97ca-40a4-8086-69473669d5b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1140.430547] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 50854a4e-8e4e-431b-8220-01bf8906b1c0] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1140.677205] env[63202]: DEBUG oslo_concurrency.lockutils [None req-1c60580d-b2ad-40ca-bb8b-827c4458e4c9 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "214c63b6-35e3-46e1-a9de-f3cea251c3c9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.326s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1140.706126] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Releasing lock "refresh_cache-9646810e-06d2-4de0-8f0c-e8719541da53" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1140.732081] env[63202]: DEBUG nova.virt.hardware [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='00119efb3f1eb9d41cca7a734a6f46e1',container_format='bare',created_at=2024-10-31T10:31:37Z,direct_url=,disk_format='vmdk',id=06831fb6-0dd7-4e39-a9d1-f69598985fd2,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-1618684641-shelved',owner='433b1d11832147b886bed7a3a4952768',properties=ImageMetaProps,protected=,size=31659008,status='active',tags=,updated_at=2024-10-31T10:31:51Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1140.732354] env[63202]: DEBUG nova.virt.hardware [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1140.732596] env[63202]: DEBUG nova.virt.hardware [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1140.732799] env[63202]: DEBUG nova.virt.hardware [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1140.732951] env[63202]: DEBUG nova.virt.hardware [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1140.733122] env[63202]: DEBUG nova.virt.hardware [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1140.733336] env[63202]: DEBUG nova.virt.hardware [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1140.733501] env[63202]: DEBUG nova.virt.hardware [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1140.733670] env[63202]: DEBUG nova.virt.hardware [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1140.733834] env[63202]: DEBUG nova.virt.hardware [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1140.734033] env[63202]: DEBUG nova.virt.hardware [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1140.734879] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5972a8c9-a929-4fc8-992c-88807e89c401 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.742976] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cefe3fd-5a95-47fd-a3c6-db1620932a77 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.756005] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cc:19:9e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e4c8c8fd-baca-4e60-97dc-ff0418d63215', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '55537461-97ca-40a4-8086-69473669d5b6', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1140.763319] env[63202]: DEBUG oslo.service.loopingcall [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1140.763834] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1140.764061] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0eca79b2-9ed5-4e8e-8a82-07a64514bea4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.785525] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1140.785525] env[63202]: value = "task-1385977" [ 1140.785525] env[63202]: _type = "Task" [ 1140.785525] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.792661] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385977, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.934097] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 5099e2b4-dcf6-4d5b-a9fc-a4d92786f52a] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1141.294980] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385977, 'name': CreateVM_Task, 'duration_secs': 0.318332} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.295954] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1141.296232] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/06831fb6-0dd7-4e39-a9d1-f69598985fd2" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1141.296349] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Acquired lock "[datastore1] devstack-image-cache_base/06831fb6-0dd7-4e39-a9d1-f69598985fd2" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1141.296756] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/06831fb6-0dd7-4e39-a9d1-f69598985fd2" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1141.297084] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-50ad3900-ab05-4af4-a4d1-0df953cf9f05 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.301647] env[63202]: DEBUG oslo_vmware.api [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Waiting for the task: (returnval){ [ 1141.301647] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]523b069a-c1ce-2a7e-c996-baf49ef727d7" [ 1141.301647] env[63202]: _type = "Task" [ 1141.301647] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.309563] env[63202]: DEBUG oslo_vmware.api [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]523b069a-c1ce-2a7e-c996-baf49ef727d7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.437385] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: a800fa5d-46a1-4a96-8ea2-26e34b869aa6] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1141.454267] env[63202]: DEBUG nova.compute.manager [req-83bde9fd-1118-4167-a52b-adfb20e8a292 req-56433f34-127b-408d-a406-fd3073a3c36c service nova] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Received event network-changed-55537461-97ca-40a4-8086-69473669d5b6 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1141.454456] env[63202]: DEBUG nova.compute.manager [req-83bde9fd-1118-4167-a52b-adfb20e8a292 req-56433f34-127b-408d-a406-fd3073a3c36c service nova] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Refreshing instance network info cache due to event network-changed-55537461-97ca-40a4-8086-69473669d5b6. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1141.454570] env[63202]: DEBUG oslo_concurrency.lockutils [req-83bde9fd-1118-4167-a52b-adfb20e8a292 req-56433f34-127b-408d-a406-fd3073a3c36c service nova] Acquiring lock "refresh_cache-9646810e-06d2-4de0-8f0c-e8719541da53" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1141.454673] env[63202]: DEBUG oslo_concurrency.lockutils [req-83bde9fd-1118-4167-a52b-adfb20e8a292 req-56433f34-127b-408d-a406-fd3073a3c36c service nova] Acquired lock "refresh_cache-9646810e-06d2-4de0-8f0c-e8719541da53" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1141.454835] env[63202]: DEBUG nova.network.neutron [req-83bde9fd-1118-4167-a52b-adfb20e8a292 req-56433f34-127b-408d-a406-fd3073a3c36c service nova] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Refreshing network info cache for port 55537461-97ca-40a4-8086-69473669d5b6 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1141.812941] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Releasing lock "[datastore1] devstack-image-cache_base/06831fb6-0dd7-4e39-a9d1-f69598985fd2" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1141.813160] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Processing image 06831fb6-0dd7-4e39-a9d1-f69598985fd2 {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1141.813397] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/06831fb6-0dd7-4e39-a9d1-f69598985fd2/06831fb6-0dd7-4e39-a9d1-f69598985fd2.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1141.813549] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Acquired lock "[datastore1] devstack-image-cache_base/06831fb6-0dd7-4e39-a9d1-f69598985fd2/06831fb6-0dd7-4e39-a9d1-f69598985fd2.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1141.813729] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1141.813976] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a5467b2a-2100-43f5-bce0-af0c9cd4d96d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.825804] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1141.825984] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1141.826691] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-98e291f4-86f7-42cf-a545-1f0379f15374 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.831405] env[63202]: DEBUG oslo_vmware.api [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Waiting for the task: (returnval){ [ 1141.831405] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52a76e0d-3d40-c351-33ef-bd6a3eb8fe8b" [ 1141.831405] env[63202]: _type = "Task" [ 1141.831405] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.839080] env[63202]: DEBUG oslo_vmware.api [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52a76e0d-3d40-c351-33ef-bd6a3eb8fe8b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.940904] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 0bec3b49-1f7f-40fe-af3c-ff87cd2a8fcb] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1142.148351] env[63202]: DEBUG nova.network.neutron [req-83bde9fd-1118-4167-a52b-adfb20e8a292 req-56433f34-127b-408d-a406-fd3073a3c36c service nova] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Updated VIF entry in instance network info cache for port 55537461-97ca-40a4-8086-69473669d5b6. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1142.148694] env[63202]: DEBUG nova.network.neutron [req-83bde9fd-1118-4167-a52b-adfb20e8a292 req-56433f34-127b-408d-a406-fd3073a3c36c service nova] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Updating instance_info_cache with network_info: [{"id": "55537461-97ca-40a4-8086-69473669d5b6", "address": "fa:16:3e:cc:19:9e", "network": {"id": "f105b723-b2a9-401d-a936-e31b4b6609f6", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-298435423-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "433b1d11832147b886bed7a3a4952768", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e4c8c8fd-baca-4e60-97dc-ff0418d63215", "external-id": "nsx-vlan-transportzone-178", "segmentation_id": 178, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap55537461-97", "ovs_interfaceid": "55537461-97ca-40a4-8086-69473669d5b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1142.150023] env[63202]: DEBUG oslo_concurrency.lockutils [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquiring lock "0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1142.150230] env[63202]: DEBUG oslo_concurrency.lockutils [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1142.341476] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Preparing fetch location {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1142.341687] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Fetch image to [datastore1] OSTACK_IMG_ff59fce9-4c39-4356-93ca-b363c6c5369c/OSTACK_IMG_ff59fce9-4c39-4356-93ca-b363c6c5369c.vmdk {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1142.341876] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Downloading stream optimized image 06831fb6-0dd7-4e39-a9d1-f69598985fd2 to [datastore1] OSTACK_IMG_ff59fce9-4c39-4356-93ca-b363c6c5369c/OSTACK_IMG_ff59fce9-4c39-4356-93ca-b363c6c5369c.vmdk on the data store datastore1 as vApp {{(pid=63202) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1142.342064] env[63202]: DEBUG nova.virt.vmwareapi.images [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Downloading image file data 06831fb6-0dd7-4e39-a9d1-f69598985fd2 to the ESX as VM named 'OSTACK_IMG_ff59fce9-4c39-4356-93ca-b363c6c5369c' {{(pid=63202) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1142.410012] env[63202]: DEBUG oslo_vmware.rw_handles [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1142.410012] env[63202]: value = "resgroup-9" [ 1142.410012] env[63202]: _type = "ResourcePool" [ 1142.410012] env[63202]: }. {{(pid=63202) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1142.410654] env[63202]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-7025be97-5795-43dd-a572-0a6a693b6613 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.431538] env[63202]: DEBUG oslo_vmware.rw_handles [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Lease: (returnval){ [ 1142.431538] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]524ace06-e0c6-1ddf-08e4-7b3a5c1f42c6" [ 1142.431538] env[63202]: _type = "HttpNfcLease" [ 1142.431538] env[63202]: } obtained for vApp import into resource pool (val){ [ 1142.431538] env[63202]: value = "resgroup-9" [ 1142.431538] env[63202]: _type = "ResourcePool" [ 1142.431538] env[63202]: }. {{(pid=63202) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1142.431820] env[63202]: DEBUG oslo_vmware.api [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Waiting for the lease: (returnval){ [ 1142.431820] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]524ace06-e0c6-1ddf-08e4-7b3a5c1f42c6" [ 1142.431820] env[63202]: _type = "HttpNfcLease" [ 1142.431820] env[63202]: } to be ready. {{(pid=63202) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1142.437484] env[63202]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1142.437484] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]524ace06-e0c6-1ddf-08e4-7b3a5c1f42c6" [ 1142.437484] env[63202]: _type = "HttpNfcLease" [ 1142.437484] env[63202]: } is initializing. {{(pid=63202) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1142.443970] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: cedc3a06-2123-4c5e-a6c3-599a3efc3c65] Instance has had 0 of 5 cleanup attempts {{(pid=63202) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1142.652607] env[63202]: DEBUG oslo_concurrency.lockutils [req-83bde9fd-1118-4167-a52b-adfb20e8a292 req-56433f34-127b-408d-a406-fd3073a3c36c service nova] Releasing lock "refresh_cache-9646810e-06d2-4de0-8f0c-e8719541da53" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1142.653111] env[63202]: DEBUG nova.compute.manager [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Starting instance... {{(pid=63202) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1142.939666] env[63202]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1142.939666] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]524ace06-e0c6-1ddf-08e4-7b3a5c1f42c6" [ 1142.939666] env[63202]: _type = "HttpNfcLease" [ 1142.939666] env[63202]: } is ready. {{(pid=63202) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1142.939944] env[63202]: DEBUG oslo_vmware.rw_handles [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1142.939944] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]524ace06-e0c6-1ddf-08e4-7b3a5c1f42c6" [ 1142.939944] env[63202]: _type = "HttpNfcLease" [ 1142.939944] env[63202]: }. {{(pid=63202) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1142.940681] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de055190-3dd9-4a6e-9f59-ea59445d8918 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.947441] env[63202]: DEBUG oslo_vmware.rw_handles [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a7841d-e2ec-e487-7704-9e1a73e77f17/disk-0.vmdk from lease info. {{(pid=63202) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1142.947714] env[63202]: DEBUG oslo_vmware.rw_handles [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Creating HTTP connection to write to file with size = 31659008 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a7841d-e2ec-e487-7704-9e1a73e77f17/disk-0.vmdk. {{(pid=63202) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1142.948853] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1142.949047] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Cleaning up deleted instances with incomplete migration {{(pid=63202) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 1143.011337] env[63202]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-ed389a4c-7976-470e-92ba-bb0b48291403 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.178515] env[63202]: DEBUG oslo_concurrency.lockutils [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1143.178791] env[63202]: DEBUG oslo_concurrency.lockutils [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1143.180430] env[63202]: INFO nova.compute.claims [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1143.451544] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1144.023360] env[63202]: DEBUG oslo_vmware.rw_handles [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Completed reading data from the image iterator. {{(pid=63202) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1144.023763] env[63202]: DEBUG oslo_vmware.rw_handles [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a7841d-e2ec-e487-7704-9e1a73e77f17/disk-0.vmdk. {{(pid=63202) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1144.024626] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0d8cd9c-994a-428f-b0bc-f28fe3fca16b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.031955] env[63202]: DEBUG oslo_vmware.rw_handles [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a7841d-e2ec-e487-7704-9e1a73e77f17/disk-0.vmdk is in state: ready. {{(pid=63202) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1144.032140] env[63202]: DEBUG oslo_vmware.rw_handles [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a7841d-e2ec-e487-7704-9e1a73e77f17/disk-0.vmdk. {{(pid=63202) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1144.032385] env[63202]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-e83211fe-b245-4203-bc48-e104fc310d45 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.243603] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4a665ce-8744-4610-a6bb-8a4e43eaa153 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.252024] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afe64ede-9915-4ff0-a541-bfbb4b245568 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.283111] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb51009a-3d04-4d85-a736-dc2b62653500 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.291020] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8b63644-b579-41a8-910b-cddae9cb171e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.305511] env[63202]: DEBUG nova.compute.provider_tree [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1144.320480] env[63202]: DEBUG oslo_vmware.rw_handles [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a7841d-e2ec-e487-7704-9e1a73e77f17/disk-0.vmdk. {{(pid=63202) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1144.320715] env[63202]: INFO nova.virt.vmwareapi.images [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Downloaded image file data 06831fb6-0dd7-4e39-a9d1-f69598985fd2 [ 1144.321522] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa1243fd-9f07-40cf-8f6f-7784f257fee7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.337211] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2c99e42e-3197-4d27-87aa-dd2d9813f530 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.364111] env[63202]: INFO nova.virt.vmwareapi.images [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] The imported VM was unregistered [ 1144.366354] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Caching image {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1144.366593] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Creating directory with path [datastore1] devstack-image-cache_base/06831fb6-0dd7-4e39-a9d1-f69598985fd2 {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1144.366838] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c4cd18db-d173-4de2-a130-d0738e286cac {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.387452] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Created directory with path [datastore1] devstack-image-cache_base/06831fb6-0dd7-4e39-a9d1-f69598985fd2 {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1144.387645] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_ff59fce9-4c39-4356-93ca-b363c6c5369c/OSTACK_IMG_ff59fce9-4c39-4356-93ca-b363c6c5369c.vmdk to [datastore1] devstack-image-cache_base/06831fb6-0dd7-4e39-a9d1-f69598985fd2/06831fb6-0dd7-4e39-a9d1-f69598985fd2.vmdk. {{(pid=63202) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1144.387887] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-f2764458-ba29-4f28-97b9-ea41e7fa5ec8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.395991] env[63202]: DEBUG oslo_vmware.api [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Waiting for the task: (returnval){ [ 1144.395991] env[63202]: value = "task-1385980" [ 1144.395991] env[63202]: _type = "Task" [ 1144.395991] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.402940] env[63202]: DEBUG oslo_vmware.api [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385980, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.808986] env[63202]: DEBUG nova.scheduler.client.report [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1144.908464] env[63202]: DEBUG oslo_vmware.api [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385980, 'name': MoveVirtualDisk_Task} progress is 18%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.314447] env[63202]: DEBUG oslo_concurrency.lockutils [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.135s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1145.315316] env[63202]: DEBUG nova.compute.manager [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Start building networks asynchronously for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1145.410190] env[63202]: DEBUG oslo_vmware.api [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385980, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.822482] env[63202]: DEBUG nova.compute.utils [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1145.824499] env[63202]: DEBUG nova.compute.manager [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Allocating IP information in the background. {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1145.824785] env[63202]: DEBUG nova.network.neutron [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] allocate_for_instance() {{(pid=63202) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1145.871088] env[63202]: DEBUG nova.policy [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '63a29c785cf240a7b6418dadc7119574', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f7d28fea097541adadf0839940568409', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63202) authorize /opt/stack/nova/nova/policy.py:201}} [ 1145.914236] env[63202]: DEBUG oslo_vmware.api [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385980, 'name': MoveVirtualDisk_Task} progress is 60%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.947985] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1145.948245] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1146.163422] env[63202]: DEBUG nova.network.neutron [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Successfully created port: 5a9c7dc7-9386-4f78-89ab-720b71c72161 {{(pid=63202) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1146.327774] env[63202]: DEBUG nova.compute.manager [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Start building block device mappings for instance. {{(pid=63202) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1146.412282] env[63202]: DEBUG oslo_vmware.api [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385980, 'name': MoveVirtualDisk_Task} progress is 80%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.459408] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1146.459408] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Starting heal instance info cache {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1146.910188] env[63202]: DEBUG oslo_vmware.api [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385980, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.339055] env[63202]: DEBUG nova.compute.manager [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Start spawning the instance on the hypervisor. {{(pid=63202) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1147.362333] env[63202]: DEBUG nova.virt.hardware [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T10:20:35Z,direct_url=,disk_format='vmdk',id=bb172aa4-a5a8-4395-9793-2416c30721cd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='980e19d195ad46dbb597408ec0afcb9e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T10:20:36Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1147.362627] env[63202]: DEBUG nova.virt.hardware [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1147.362804] env[63202]: DEBUG nova.virt.hardware [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1147.362994] env[63202]: DEBUG nova.virt.hardware [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1147.363162] env[63202]: DEBUG nova.virt.hardware [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1147.363313] env[63202]: DEBUG nova.virt.hardware [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1147.363518] env[63202]: DEBUG nova.virt.hardware [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1147.363681] env[63202]: DEBUG nova.virt.hardware [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1147.363848] env[63202]: DEBUG nova.virt.hardware [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1147.364070] env[63202]: DEBUG nova.virt.hardware [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1147.364205] env[63202]: DEBUG nova.virt.hardware [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1147.365085] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-386e63c9-5b0c-4133-b69b-a9ac6df80cd6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.372972] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76b0c6b9-c56f-4468-8394-0c72e993e87c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.408555] env[63202]: DEBUG oslo_vmware.api [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385980, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.536724} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.408798] env[63202]: INFO nova.virt.vmwareapi.ds_util [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_ff59fce9-4c39-4356-93ca-b363c6c5369c/OSTACK_IMG_ff59fce9-4c39-4356-93ca-b363c6c5369c.vmdk to [datastore1] devstack-image-cache_base/06831fb6-0dd7-4e39-a9d1-f69598985fd2/06831fb6-0dd7-4e39-a9d1-f69598985fd2.vmdk. [ 1147.408984] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Cleaning up location [datastore1] OSTACK_IMG_ff59fce9-4c39-4356-93ca-b363c6c5369c {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1147.409167] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_ff59fce9-4c39-4356-93ca-b363c6c5369c {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1147.409408] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f19e1744-a5de-4b5b-89d6-afa1f9e8832c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.415460] env[63202]: DEBUG oslo_vmware.api [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Waiting for the task: (returnval){ [ 1147.415460] env[63202]: value = "task-1385981" [ 1147.415460] env[63202]: _type = "Task" [ 1147.415460] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.422601] env[63202]: DEBUG oslo_vmware.api [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385981, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.464313] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Didn't find any instances for network info cache update. {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1147.464541] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1147.464703] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1147.464848] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1147.464999] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._sync_power_states {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1147.582387] env[63202]: DEBUG nova.compute.manager [req-3094881d-c45d-4fb2-9883-844bd0f0a803 req-44869353-e422-4965-9fa5-8e0d4fc2efe0 service nova] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Received event network-vif-plugged-5a9c7dc7-9386-4f78-89ab-720b71c72161 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1147.582632] env[63202]: DEBUG oslo_concurrency.lockutils [req-3094881d-c45d-4fb2-9883-844bd0f0a803 req-44869353-e422-4965-9fa5-8e0d4fc2efe0 service nova] Acquiring lock "0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1147.582842] env[63202]: DEBUG oslo_concurrency.lockutils [req-3094881d-c45d-4fb2-9883-844bd0f0a803 req-44869353-e422-4965-9fa5-8e0d4fc2efe0 service nova] Lock "0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1147.583048] env[63202]: DEBUG oslo_concurrency.lockutils [req-3094881d-c45d-4fb2-9883-844bd0f0a803 req-44869353-e422-4965-9fa5-8e0d4fc2efe0 service nova] Lock "0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1147.583226] env[63202]: DEBUG nova.compute.manager [req-3094881d-c45d-4fb2-9883-844bd0f0a803 req-44869353-e422-4965-9fa5-8e0d4fc2efe0 service nova] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] No waiting events found dispatching network-vif-plugged-5a9c7dc7-9386-4f78-89ab-720b71c72161 {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1147.583390] env[63202]: WARNING nova.compute.manager [req-3094881d-c45d-4fb2-9883-844bd0f0a803 req-44869353-e422-4965-9fa5-8e0d4fc2efe0 service nova] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Received unexpected event network-vif-plugged-5a9c7dc7-9386-4f78-89ab-720b71c72161 for instance with vm_state building and task_state spawning. [ 1147.654125] env[63202]: DEBUG nova.network.neutron [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Successfully updated port: 5a9c7dc7-9386-4f78-89ab-720b71c72161 {{(pid=63202) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1147.926500] env[63202]: DEBUG oslo_vmware.api [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385981, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.034603} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.926910] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1147.927211] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Releasing lock "[datastore1] devstack-image-cache_base/06831fb6-0dd7-4e39-a9d1-f69598985fd2/06831fb6-0dd7-4e39-a9d1-f69598985fd2.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1147.927515] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/06831fb6-0dd7-4e39-a9d1-f69598985fd2/06831fb6-0dd7-4e39-a9d1-f69598985fd2.vmdk to [datastore1] 9646810e-06d2-4de0-8f0c-e8719541da53/9646810e-06d2-4de0-8f0c-e8719541da53.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1147.927775] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-60aa0828-35c2-4ba7-a69d-8e38a1b98288 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.934603] env[63202]: DEBUG oslo_vmware.api [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Waiting for the task: (returnval){ [ 1147.934603] env[63202]: value = "task-1385982" [ 1147.934603] env[63202]: _type = "Task" [ 1147.934603] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.942388] env[63202]: DEBUG oslo_vmware.api [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385982, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.968704] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Getting list of instances from cluster (obj){ [ 1147.968704] env[63202]: value = "domain-c8" [ 1147.968704] env[63202]: _type = "ClusterComputeResource" [ 1147.968704] env[63202]: } {{(pid=63202) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1147.969682] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19ff8e9e-d9dc-423e-bea0-e6c3ab1d0d2d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.980165] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Got total of 2 instances {{(pid=63202) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1147.980314] env[63202]: WARNING nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] While synchronizing instance power states, found 3 instances in the database and 2 instances on the hypervisor. [ 1147.980476] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Triggering sync for uuid 9646810e-06d2-4de0-8f0c-e8719541da53 {{(pid=63202) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1147.980674] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Triggering sync for uuid ed60fb5e-b08f-4846-942d-c4edde39d1ae {{(pid=63202) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1147.980829] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Triggering sync for uuid 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8 {{(pid=63202) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1147.981158] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Acquiring lock "9646810e-06d2-4de0-8f0c-e8719541da53" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1147.981398] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Acquiring lock "ed60fb5e-b08f-4846-942d-c4edde39d1ae" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1147.981593] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "ed60fb5e-b08f-4846-942d-c4edde39d1ae" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1147.981886] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Acquiring lock "0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1147.982089] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1147.982222] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63202) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1147.982943] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c76dac7-dfe5-4f8d-9fd4-f7521f394240 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.985322] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager.update_available_resource {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1148.156241] env[63202]: DEBUG oslo_concurrency.lockutils [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquiring lock "refresh_cache-0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1148.156438] env[63202]: DEBUG oslo_concurrency.lockutils [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquired lock "refresh_cache-0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1148.156999] env[63202]: DEBUG nova.network.neutron [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1148.447814] env[63202]: DEBUG oslo_vmware.api [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385982, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.488082] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1148.488363] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1148.488488] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1148.488638] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63202) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1148.489724] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20cd65c5-3111-45bd-8414-fdec2a56dbcf {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.492760] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "ed60fb5e-b08f-4846-942d-c4edde39d1ae" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.511s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1148.500787] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed0805a1-5715-48ef-b88c-306372979f8a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.520469] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d039893-72f7-487a-b2d1-8613ca685db6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.530343] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b87a48dd-60f3-4b84-abae-fde7f4ef6da4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.563350] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181025MB free_disk=176GB free_vcpus=48 pci_devices=None {{(pid=63202) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1148.563516] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1148.563733] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1148.690136] env[63202]: DEBUG nova.network.neutron [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Instance cache missing network info. {{(pid=63202) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1148.839437] env[63202]: DEBUG nova.network.neutron [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Updating instance_info_cache with network_info: [{"id": "5a9c7dc7-9386-4f78-89ab-720b71c72161", "address": "fa:16:3e:21:0d:71", "network": {"id": "0aa55fee-953e-4c2d-b88a-43b0cf86c7f3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-878071414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7d28fea097541adadf0839940568409", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5a9c7dc7-93", "ovs_interfaceid": "5a9c7dc7-9386-4f78-89ab-720b71c72161", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1148.947813] env[63202]: DEBUG oslo_vmware.api [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385982, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.343023] env[63202]: DEBUG oslo_concurrency.lockutils [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Releasing lock "refresh_cache-0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1149.343023] env[63202]: DEBUG nova.compute.manager [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Instance network_info: |[{"id": "5a9c7dc7-9386-4f78-89ab-720b71c72161", "address": "fa:16:3e:21:0d:71", "network": {"id": "0aa55fee-953e-4c2d-b88a-43b0cf86c7f3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-878071414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7d28fea097541adadf0839940568409", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5a9c7dc7-93", "ovs_interfaceid": "5a9c7dc7-9386-4f78-89ab-720b71c72161", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63202) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1149.343306] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:21:0d:71', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5a9c7dc7-9386-4f78-89ab-720b71c72161', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1149.350950] env[63202]: DEBUG oslo.service.loopingcall [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1149.351225] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1149.351464] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-66ed0bc7-3826-48f0-bb27-c8ff4ea6133a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.375754] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1149.375754] env[63202]: value = "task-1385983" [ 1149.375754] env[63202]: _type = "Task" [ 1149.375754] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.388577] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385983, 'name': CreateVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.449053] env[63202]: DEBUG oslo_vmware.api [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385982, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.590715] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance ed60fb5e-b08f-4846-942d-c4edde39d1ae actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1149.590870] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 9646810e-06d2-4de0-8f0c-e8719541da53 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1149.590995] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1149.591191] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=63202) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1149.591333] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=63202) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1149.608508] env[63202]: DEBUG nova.compute.manager [req-1832fdaa-ca6a-4444-be26-c66c1f3dd1b6 req-10b9f206-6c6b-4d0c-b2a8-add3b9a75200 service nova] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Received event network-changed-5a9c7dc7-9386-4f78-89ab-720b71c72161 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1149.608762] env[63202]: DEBUG nova.compute.manager [req-1832fdaa-ca6a-4444-be26-c66c1f3dd1b6 req-10b9f206-6c6b-4d0c-b2a8-add3b9a75200 service nova] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Refreshing instance network info cache due to event network-changed-5a9c7dc7-9386-4f78-89ab-720b71c72161. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1149.609028] env[63202]: DEBUG oslo_concurrency.lockutils [req-1832fdaa-ca6a-4444-be26-c66c1f3dd1b6 req-10b9f206-6c6b-4d0c-b2a8-add3b9a75200 service nova] Acquiring lock "refresh_cache-0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1149.609206] env[63202]: DEBUG oslo_concurrency.lockutils [req-1832fdaa-ca6a-4444-be26-c66c1f3dd1b6 req-10b9f206-6c6b-4d0c-b2a8-add3b9a75200 service nova] Acquired lock "refresh_cache-0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1149.609398] env[63202]: DEBUG nova.network.neutron [req-1832fdaa-ca6a-4444-be26-c66c1f3dd1b6 req-10b9f206-6c6b-4d0c-b2a8-add3b9a75200 service nova] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Refreshing network info cache for port 5a9c7dc7-9386-4f78-89ab-720b71c72161 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1149.650461] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aea42e96-989e-4eba-b463-d1c3f6f6487a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.664071] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74d9dc66-438a-4178-9869-3423da5d75b0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.700057] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b123b21a-3ddc-4c3d-8630-ee72d8a80cea {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.710765] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d77449fa-8b81-465e-8c99-a318810488ce {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.728232] env[63202]: DEBUG nova.compute.provider_tree [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1149.890327] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385983, 'name': CreateVM_Task} progress is 25%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.947273] env[63202]: DEBUG oslo_vmware.api [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385982, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.231484] env[63202]: DEBUG nova.scheduler.client.report [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1150.321660] env[63202]: DEBUG nova.network.neutron [req-1832fdaa-ca6a-4444-be26-c66c1f3dd1b6 req-10b9f206-6c6b-4d0c-b2a8-add3b9a75200 service nova] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Updated VIF entry in instance network info cache for port 5a9c7dc7-9386-4f78-89ab-720b71c72161. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1150.322038] env[63202]: DEBUG nova.network.neutron [req-1832fdaa-ca6a-4444-be26-c66c1f3dd1b6 req-10b9f206-6c6b-4d0c-b2a8-add3b9a75200 service nova] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Updating instance_info_cache with network_info: [{"id": "5a9c7dc7-9386-4f78-89ab-720b71c72161", "address": "fa:16:3e:21:0d:71", "network": {"id": "0aa55fee-953e-4c2d-b88a-43b0cf86c7f3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-878071414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7d28fea097541adadf0839940568409", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5a9c7dc7-93", "ovs_interfaceid": "5a9c7dc7-9386-4f78-89ab-720b71c72161", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1150.386729] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385983, 'name': CreateVM_Task} progress is 99%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.446947] env[63202]: DEBUG oslo_vmware.api [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385982, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.416722} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.447226] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/06831fb6-0dd7-4e39-a9d1-f69598985fd2/06831fb6-0dd7-4e39-a9d1-f69598985fd2.vmdk to [datastore1] 9646810e-06d2-4de0-8f0c-e8719541da53/9646810e-06d2-4de0-8f0c-e8719541da53.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1150.448020] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d0b9f51-3474-40e3-9e62-214fe5c1e2cb {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.469829] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] 9646810e-06d2-4de0-8f0c-e8719541da53/9646810e-06d2-4de0-8f0c-e8719541da53.vmdk or device None with type streamOptimized {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1150.470401] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c3c5524c-e042-4a8b-ab4c-44f43c612dd1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.489342] env[63202]: DEBUG oslo_vmware.api [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Waiting for the task: (returnval){ [ 1150.489342] env[63202]: value = "task-1385984" [ 1150.489342] env[63202]: _type = "Task" [ 1150.489342] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.496734] env[63202]: DEBUG oslo_vmware.api [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385984, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.736998] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63202) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1150.737188] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.173s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1150.825224] env[63202]: DEBUG oslo_concurrency.lockutils [req-1832fdaa-ca6a-4444-be26-c66c1f3dd1b6 req-10b9f206-6c6b-4d0c-b2a8-add3b9a75200 service nova] Releasing lock "refresh_cache-0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1150.887097] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1385983, 'name': CreateVM_Task, 'duration_secs': 1.451958} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.887257] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1150.887945] env[63202]: DEBUG oslo_concurrency.lockutils [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1150.888129] env[63202]: DEBUG oslo_concurrency.lockutils [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1150.888459] env[63202]: DEBUG oslo_concurrency.lockutils [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1150.888726] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-130c52ad-86af-4a98-b920-4d00e02d6ca8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.892905] env[63202]: DEBUG oslo_vmware.api [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 1150.892905] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]523cbd53-7fb9-adf8-d7ec-7ce37783bbf9" [ 1150.892905] env[63202]: _type = "Task" [ 1150.892905] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.900130] env[63202]: DEBUG oslo_vmware.api [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]523cbd53-7fb9-adf8-d7ec-7ce37783bbf9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.998652] env[63202]: DEBUG oslo_vmware.api [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385984, 'name': ReconfigVM_Task, 'duration_secs': 0.269703} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.998876] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Reconfigured VM instance instance-00000066 to attach disk [datastore1] 9646810e-06d2-4de0-8f0c-e8719541da53/9646810e-06d2-4de0-8f0c-e8719541da53.vmdk or device None with type streamOptimized {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1150.999507] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a046ebb7-b634-4aa8-9b46-c466682cda7e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.004759] env[63202]: DEBUG oslo_vmware.api [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Waiting for the task: (returnval){ [ 1151.004759] env[63202]: value = "task-1385985" [ 1151.004759] env[63202]: _type = "Task" [ 1151.004759] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.011807] env[63202]: DEBUG oslo_vmware.api [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385985, 'name': Rename_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.404492] env[63202]: DEBUG oslo_vmware.api [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]523cbd53-7fb9-adf8-d7ec-7ce37783bbf9, 'name': SearchDatastore_Task, 'duration_secs': 0.009073} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.404731] env[63202]: DEBUG oslo_concurrency.lockutils [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1151.404963] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Processing image bb172aa4-a5a8-4395-9793-2416c30721cd {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1151.405215] env[63202]: DEBUG oslo_concurrency.lockutils [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1151.405365] env[63202]: DEBUG oslo_concurrency.lockutils [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1151.405543] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1151.405799] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6951975b-72d7-4e8c-856d-a935ae5dce1e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.413322] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1151.413490] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1151.414191] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-336012a3-30b2-4970-a5e5-e04d1da91fd0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.418887] env[63202]: DEBUG oslo_vmware.api [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 1151.418887] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]5201153a-fc72-bc39-17a1-176ca1ad8420" [ 1151.418887] env[63202]: _type = "Task" [ 1151.418887] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.426313] env[63202]: DEBUG oslo_vmware.api [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]5201153a-fc72-bc39-17a1-176ca1ad8420, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.513962] env[63202]: DEBUG oslo_vmware.api [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385985, 'name': Rename_Task, 'duration_secs': 0.132762} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.514247] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1151.514473] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1cc9d829-7b6e-4f03-999d-7b844e65dbd7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.520793] env[63202]: DEBUG oslo_vmware.api [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Waiting for the task: (returnval){ [ 1151.520793] env[63202]: value = "task-1385986" [ 1151.520793] env[63202]: _type = "Task" [ 1151.520793] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.528103] env[63202]: DEBUG oslo_vmware.api [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385986, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.929021] env[63202]: DEBUG oslo_vmware.api [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]5201153a-fc72-bc39-17a1-176ca1ad8420, 'name': SearchDatastore_Task, 'duration_secs': 0.008215} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.929782] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e8fdf475-4f62-40d5-b6cb-429461b0073f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.934715] env[63202]: DEBUG oslo_vmware.api [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 1151.934715] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52773127-8dbd-e70f-d35b-d45a06633bcf" [ 1151.934715] env[63202]: _type = "Task" [ 1151.934715] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.942218] env[63202]: DEBUG oslo_vmware.api [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52773127-8dbd-e70f-d35b-d45a06633bcf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.030163] env[63202]: DEBUG oslo_vmware.api [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385986, 'name': PowerOnVM_Task, 'duration_secs': 0.463215} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.030428] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1152.136878] env[63202]: DEBUG nova.compute.manager [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1152.137859] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38e458e1-bed4-4952-815a-1a2367c07514 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.446546] env[63202]: DEBUG oslo_vmware.api [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52773127-8dbd-e70f-d35b-d45a06633bcf, 'name': SearchDatastore_Task, 'duration_secs': 0.009105} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.446824] env[63202]: DEBUG oslo_concurrency.lockutils [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1152.447095] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8/0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1152.447361] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-18168a1b-40f0-479b-9ab1-b02bbe9a2ec1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.454453] env[63202]: DEBUG oslo_vmware.api [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 1152.454453] env[63202]: value = "task-1385987" [ 1152.454453] env[63202]: _type = "Task" [ 1152.454453] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.462411] env[63202]: DEBUG oslo_vmware.api [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385987, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.658189] env[63202]: DEBUG oslo_concurrency.lockutils [None req-a354702e-c303-4467-a85e-28bd02c05892 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Lock "9646810e-06d2-4de0-8f0c-e8719541da53" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 18.814s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1152.659315] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "9646810e-06d2-4de0-8f0c-e8719541da53" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 4.678s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1152.659530] env[63202]: INFO nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] During sync_power_state the instance has a pending task (spawning). Skip. [ 1152.660313] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "9646810e-06d2-4de0-8f0c-e8719541da53" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1152.965247] env[63202]: DEBUG oslo_vmware.api [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385987, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.48781} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.965520] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bb172aa4-a5a8-4395-9793-2416c30721cd/bb172aa4-a5a8-4395-9793-2416c30721cd.vmdk to [datastore1] 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8/0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1152.965768] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Extending root virtual disk to 1048576 {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1152.966039] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cfb34e42-eb40-495c-8fbd-a6fb81c383d0 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.971869] env[63202]: DEBUG oslo_vmware.api [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 1152.971869] env[63202]: value = "task-1385988" [ 1152.971869] env[63202]: _type = "Task" [ 1152.971869] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.978743] env[63202]: DEBUG oslo_vmware.api [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385988, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.482340] env[63202]: DEBUG oslo_vmware.api [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385988, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062383} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.482677] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Extended root virtual disk {{(pid=63202) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1153.483557] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-463446ac-93e1-4a32-a401-dadb86ace4db {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.505725] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8/0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1153.507372] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-67ee4dbd-401d-4040-9ed3-a7a801b18456 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.521916] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-621ec72e-e676-41c2-bbbf-8964f862c44b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.528802] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-cebd3d54-607f-4995-9b67-fc31ffcdbc6d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Suspending the VM {{(pid=63202) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1153.529876] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-7848525f-0e98-484f-aad1-6c32f1bbcac2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.531241] env[63202]: DEBUG oslo_vmware.api [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 1153.531241] env[63202]: value = "task-1385989" [ 1153.531241] env[63202]: _type = "Task" [ 1153.531241] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.535739] env[63202]: DEBUG oslo_vmware.api [None req-cebd3d54-607f-4995-9b67-fc31ffcdbc6d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Waiting for the task: (returnval){ [ 1153.535739] env[63202]: value = "task-1385990" [ 1153.535739] env[63202]: _type = "Task" [ 1153.535739] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.538637] env[63202]: DEBUG oslo_vmware.api [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385989, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.546051] env[63202]: DEBUG oslo_vmware.api [None req-cebd3d54-607f-4995-9b67-fc31ffcdbc6d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385990, 'name': SuspendVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.041328] env[63202]: DEBUG oslo_vmware.api [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385989, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.048321] env[63202]: DEBUG oslo_vmware.api [None req-cebd3d54-607f-4995-9b67-fc31ffcdbc6d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385990, 'name': SuspendVM_Task} progress is 70%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.542601] env[63202]: DEBUG oslo_vmware.api [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385989, 'name': ReconfigVM_Task, 'duration_secs': 0.659199} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.545365] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Reconfigured VM instance instance-0000006b to attach disk [datastore1] 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8/0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8.vmdk or device None with type sparse {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1154.545906] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-402c9eb1-0cd5-406e-8192-4a8c09132c7c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.551662] env[63202]: DEBUG oslo_vmware.api [None req-cebd3d54-607f-4995-9b67-fc31ffcdbc6d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385990, 'name': SuspendVM_Task, 'duration_secs': 0.670109} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.552729] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-cebd3d54-607f-4995-9b67-fc31ffcdbc6d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Suspended the VM {{(pid=63202) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1154.552915] env[63202]: DEBUG nova.compute.manager [None req-cebd3d54-607f-4995-9b67-fc31ffcdbc6d tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1154.553228] env[63202]: DEBUG oslo_vmware.api [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 1154.553228] env[63202]: value = "task-1385991" [ 1154.553228] env[63202]: _type = "Task" [ 1154.553228] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.554009] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec9b0503-c4d7-4a69-a71f-0f95c09a4570 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.566060] env[63202]: DEBUG oslo_vmware.api [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385991, 'name': Rename_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.065915] env[63202]: DEBUG oslo_vmware.api [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385991, 'name': Rename_Task, 'duration_secs': 0.150587} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.066308] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1155.066424] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-71408898-ec0c-4725-9b68-4e1f878ec28f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.074075] env[63202]: DEBUG oslo_vmware.api [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 1155.074075] env[63202]: value = "task-1385992" [ 1155.074075] env[63202]: _type = "Task" [ 1155.074075] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.080805] env[63202]: DEBUG oslo_vmware.api [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385992, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.583604] env[63202]: DEBUG oslo_vmware.api [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1385992, 'name': PowerOnVM_Task, 'duration_secs': 0.446859} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.583892] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1155.584113] env[63202]: INFO nova.compute.manager [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Took 8.25 seconds to spawn the instance on the hypervisor. [ 1155.584298] env[63202]: DEBUG nova.compute.manager [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1155.585056] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57dc5ad4-f050-4970-9739-1d836084935e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.856025] env[63202]: INFO nova.compute.manager [None req-b6aa701b-0f0f-4385-b05e-4b1aa4e3e404 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Resuming [ 1155.856652] env[63202]: DEBUG nova.objects.instance [None req-b6aa701b-0f0f-4385-b05e-4b1aa4e3e404 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Lazy-loading 'flavor' on Instance uuid 9646810e-06d2-4de0-8f0c-e8719541da53 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1156.102199] env[63202]: INFO nova.compute.manager [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Took 12.94 seconds to build instance. [ 1156.508985] env[63202]: DEBUG nova.compute.manager [req-fd2feb62-eae4-409f-b2ce-ca581cb9cd45 req-29b479d7-5162-4994-9898-209cb6817503 service nova] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Received event network-changed-5a9c7dc7-9386-4f78-89ab-720b71c72161 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1156.509253] env[63202]: DEBUG nova.compute.manager [req-fd2feb62-eae4-409f-b2ce-ca581cb9cd45 req-29b479d7-5162-4994-9898-209cb6817503 service nova] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Refreshing instance network info cache due to event network-changed-5a9c7dc7-9386-4f78-89ab-720b71c72161. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1156.509441] env[63202]: DEBUG oslo_concurrency.lockutils [req-fd2feb62-eae4-409f-b2ce-ca581cb9cd45 req-29b479d7-5162-4994-9898-209cb6817503 service nova] Acquiring lock "refresh_cache-0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1156.509606] env[63202]: DEBUG oslo_concurrency.lockutils [req-fd2feb62-eae4-409f-b2ce-ca581cb9cd45 req-29b479d7-5162-4994-9898-209cb6817503 service nova] Acquired lock "refresh_cache-0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1156.509769] env[63202]: DEBUG nova.network.neutron [req-fd2feb62-eae4-409f-b2ce-ca581cb9cd45 req-29b479d7-5162-4994-9898-209cb6817503 service nova] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Refreshing network info cache for port 5a9c7dc7-9386-4f78-89ab-720b71c72161 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1156.604515] env[63202]: DEBUG oslo_concurrency.lockutils [None req-68589176-5802-4189-8486-88bb594e7bbe tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.454s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1156.604820] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 8.623s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1156.605041] env[63202]: INFO nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] During sync_power_state the instance has a pending task (spawning). Skip. [ 1156.605230] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1156.864805] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b6aa701b-0f0f-4385-b05e-4b1aa4e3e404 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Acquiring lock "refresh_cache-9646810e-06d2-4de0-8f0c-e8719541da53" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1156.865016] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b6aa701b-0f0f-4385-b05e-4b1aa4e3e404 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Acquired lock "refresh_cache-9646810e-06d2-4de0-8f0c-e8719541da53" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1156.865212] env[63202]: DEBUG nova.network.neutron [None req-b6aa701b-0f0f-4385-b05e-4b1aa4e3e404 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1157.213630] env[63202]: DEBUG nova.network.neutron [req-fd2feb62-eae4-409f-b2ce-ca581cb9cd45 req-29b479d7-5162-4994-9898-209cb6817503 service nova] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Updated VIF entry in instance network info cache for port 5a9c7dc7-9386-4f78-89ab-720b71c72161. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1157.214034] env[63202]: DEBUG nova.network.neutron [req-fd2feb62-eae4-409f-b2ce-ca581cb9cd45 req-29b479d7-5162-4994-9898-209cb6817503 service nova] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Updating instance_info_cache with network_info: [{"id": "5a9c7dc7-9386-4f78-89ab-720b71c72161", "address": "fa:16:3e:21:0d:71", "network": {"id": "0aa55fee-953e-4c2d-b88a-43b0cf86c7f3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-878071414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.202", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7d28fea097541adadf0839940568409", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33bcfd29-ad69-41ad-8e7f-55c1a3cf2dce", "external-id": "nsx-vlan-transportzone-725", "segmentation_id": 725, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5a9c7dc7-93", "ovs_interfaceid": "5a9c7dc7-9386-4f78-89ab-720b71c72161", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1157.546193] env[63202]: DEBUG nova.network.neutron [None req-b6aa701b-0f0f-4385-b05e-4b1aa4e3e404 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Updating instance_info_cache with network_info: [{"id": "55537461-97ca-40a4-8086-69473669d5b6", "address": "fa:16:3e:cc:19:9e", "network": {"id": "f105b723-b2a9-401d-a936-e31b4b6609f6", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-298435423-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "433b1d11832147b886bed7a3a4952768", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e4c8c8fd-baca-4e60-97dc-ff0418d63215", "external-id": "nsx-vlan-transportzone-178", "segmentation_id": 178, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap55537461-97", "ovs_interfaceid": "55537461-97ca-40a4-8086-69473669d5b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1157.716656] env[63202]: DEBUG oslo_concurrency.lockutils [req-fd2feb62-eae4-409f-b2ce-ca581cb9cd45 req-29b479d7-5162-4994-9898-209cb6817503 service nova] Releasing lock "refresh_cache-0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1158.048606] env[63202]: DEBUG oslo_concurrency.lockutils [None req-b6aa701b-0f0f-4385-b05e-4b1aa4e3e404 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Releasing lock "refresh_cache-9646810e-06d2-4de0-8f0c-e8719541da53" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1158.049601] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4b30f2f-041d-414c-b987-b50e4497cc28 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.056159] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-b6aa701b-0f0f-4385-b05e-4b1aa4e3e404 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Resuming the VM {{(pid=63202) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 1158.056398] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8e82a755-c1ed-4382-a79b-c9fb018f3d3d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.062582] env[63202]: DEBUG oslo_vmware.api [None req-b6aa701b-0f0f-4385-b05e-4b1aa4e3e404 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Waiting for the task: (returnval){ [ 1158.062582] env[63202]: value = "task-1385993" [ 1158.062582] env[63202]: _type = "Task" [ 1158.062582] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.070259] env[63202]: DEBUG oslo_vmware.api [None req-b6aa701b-0f0f-4385-b05e-4b1aa4e3e404 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385993, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.577239] env[63202]: DEBUG oslo_vmware.api [None req-b6aa701b-0f0f-4385-b05e-4b1aa4e3e404 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385993, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.072495] env[63202]: DEBUG oslo_vmware.api [None req-b6aa701b-0f0f-4385-b05e-4b1aa4e3e404 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385993, 'name': PowerOnVM_Task, 'duration_secs': 0.543735} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.072797] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-b6aa701b-0f0f-4385-b05e-4b1aa4e3e404 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Resumed the VM {{(pid=63202) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 1159.072958] env[63202]: DEBUG nova.compute.manager [None req-b6aa701b-0f0f-4385-b05e-4b1aa4e3e404 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1159.073734] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33750ecb-18a9-484b-82b8-4fa3e17334a1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.089209] env[63202]: DEBUG oslo_concurrency.lockutils [None req-15c786ed-599e-44c0-9c79-17a5be09ec80 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Acquiring lock "9646810e-06d2-4de0-8f0c-e8719541da53" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1164.089519] env[63202]: DEBUG oslo_concurrency.lockutils [None req-15c786ed-599e-44c0-9c79-17a5be09ec80 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Lock "9646810e-06d2-4de0-8f0c-e8719541da53" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1164.089705] env[63202]: DEBUG oslo_concurrency.lockutils [None req-15c786ed-599e-44c0-9c79-17a5be09ec80 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Acquiring lock "9646810e-06d2-4de0-8f0c-e8719541da53-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1164.089896] env[63202]: DEBUG oslo_concurrency.lockutils [None req-15c786ed-599e-44c0-9c79-17a5be09ec80 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Lock "9646810e-06d2-4de0-8f0c-e8719541da53-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1164.090076] env[63202]: DEBUG oslo_concurrency.lockutils [None req-15c786ed-599e-44c0-9c79-17a5be09ec80 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Lock "9646810e-06d2-4de0-8f0c-e8719541da53-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1164.092219] env[63202]: INFO nova.compute.manager [None req-15c786ed-599e-44c0-9c79-17a5be09ec80 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Terminating instance [ 1164.094023] env[63202]: DEBUG nova.compute.manager [None req-15c786ed-599e-44c0-9c79-17a5be09ec80 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1164.094222] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-15c786ed-599e-44c0-9c79-17a5be09ec80 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1164.095063] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e30d518b-e3da-4acc-ba5b-c93c31e636ea {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.102720] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-15c786ed-599e-44c0-9c79-17a5be09ec80 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1164.102949] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3d33d476-6e08-4ca9-8c03-356a2897f16d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.108820] env[63202]: DEBUG oslo_vmware.api [None req-15c786ed-599e-44c0-9c79-17a5be09ec80 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Waiting for the task: (returnval){ [ 1164.108820] env[63202]: value = "task-1385994" [ 1164.108820] env[63202]: _type = "Task" [ 1164.108820] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.116882] env[63202]: DEBUG oslo_vmware.api [None req-15c786ed-599e-44c0-9c79-17a5be09ec80 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385994, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.619063] env[63202]: DEBUG oslo_vmware.api [None req-15c786ed-599e-44c0-9c79-17a5be09ec80 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385994, 'name': PowerOffVM_Task, 'duration_secs': 0.160863} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.619255] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-15c786ed-599e-44c0-9c79-17a5be09ec80 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1164.619368] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-15c786ed-599e-44c0-9c79-17a5be09ec80 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1164.619607] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2387ac65-d8f9-439a-8c6b-16c804004d7c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.708422] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-15c786ed-599e-44c0-9c79-17a5be09ec80 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1164.708674] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-15c786ed-599e-44c0-9c79-17a5be09ec80 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Deleting contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1164.708886] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-15c786ed-599e-44c0-9c79-17a5be09ec80 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Deleting the datastore file [datastore1] 9646810e-06d2-4de0-8f0c-e8719541da53 {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1164.709276] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c6d61d88-c50a-4c23-948c-8c410e66118a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.717092] env[63202]: DEBUG oslo_vmware.api [None req-15c786ed-599e-44c0-9c79-17a5be09ec80 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Waiting for the task: (returnval){ [ 1164.717092] env[63202]: value = "task-1385996" [ 1164.717092] env[63202]: _type = "Task" [ 1164.717092] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.725196] env[63202]: DEBUG oslo_vmware.api [None req-15c786ed-599e-44c0-9c79-17a5be09ec80 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385996, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.226676] env[63202]: DEBUG oslo_vmware.api [None req-15c786ed-599e-44c0-9c79-17a5be09ec80 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Task: {'id': task-1385996, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.130856} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.227049] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-15c786ed-599e-44c0-9c79-17a5be09ec80 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1165.227106] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-15c786ed-599e-44c0-9c79-17a5be09ec80 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Deleted contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1165.227272] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-15c786ed-599e-44c0-9c79-17a5be09ec80 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1165.227445] env[63202]: INFO nova.compute.manager [None req-15c786ed-599e-44c0-9c79-17a5be09ec80 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1165.227683] env[63202]: DEBUG oslo.service.loopingcall [None req-15c786ed-599e-44c0-9c79-17a5be09ec80 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1165.227875] env[63202]: DEBUG nova.compute.manager [-] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1165.227970] env[63202]: DEBUG nova.network.neutron [-] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1165.474361] env[63202]: DEBUG nova.compute.manager [req-ffeceece-e241-47ed-9170-94aafabf4892 req-29594aca-5f07-4b70-a594-54a40ec87094 service nova] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Received event network-vif-deleted-55537461-97ca-40a4-8086-69473669d5b6 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1165.474361] env[63202]: INFO nova.compute.manager [req-ffeceece-e241-47ed-9170-94aafabf4892 req-29594aca-5f07-4b70-a594-54a40ec87094 service nova] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Neutron deleted interface 55537461-97ca-40a4-8086-69473669d5b6; detaching it from the instance and deleting it from the info cache [ 1165.474361] env[63202]: DEBUG nova.network.neutron [req-ffeceece-e241-47ed-9170-94aafabf4892 req-29594aca-5f07-4b70-a594-54a40ec87094 service nova] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1165.953279] env[63202]: DEBUG nova.network.neutron [-] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1165.976475] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-33652d21-4e67-41c3-8390-e849c4902754 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.986409] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26268126-5e6a-4061-b4f5-78bc60f1698e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.009874] env[63202]: DEBUG nova.compute.manager [req-ffeceece-e241-47ed-9170-94aafabf4892 req-29594aca-5f07-4b70-a594-54a40ec87094 service nova] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Detach interface failed, port_id=55537461-97ca-40a4-8086-69473669d5b6, reason: Instance 9646810e-06d2-4de0-8f0c-e8719541da53 could not be found. {{(pid=63202) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1166.456333] env[63202]: INFO nova.compute.manager [-] [instance: 9646810e-06d2-4de0-8f0c-e8719541da53] Took 1.23 seconds to deallocate network for instance. [ 1166.963249] env[63202]: DEBUG oslo_concurrency.lockutils [None req-15c786ed-599e-44c0-9c79-17a5be09ec80 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1166.963554] env[63202]: DEBUG oslo_concurrency.lockutils [None req-15c786ed-599e-44c0-9c79-17a5be09ec80 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1166.963838] env[63202]: DEBUG nova.objects.instance [None req-15c786ed-599e-44c0-9c79-17a5be09ec80 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Lazy-loading 'resources' on Instance uuid 9646810e-06d2-4de0-8f0c-e8719541da53 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1167.518124] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-159e2728-76f4-4dbe-9989-bfb7a248eeb4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.525784] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd57e146-f7c1-4dcf-9187-0370c2ee5dec {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.555394] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c999dc6e-5db6-42eb-a280-ecc0b41d938d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.562060] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40c9d498-e9a3-49cc-9cb6-3ded368b5653 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.574401] env[63202]: DEBUG nova.compute.provider_tree [None req-15c786ed-599e-44c0-9c79-17a5be09ec80 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1168.077259] env[63202]: DEBUG nova.scheduler.client.report [None req-15c786ed-599e-44c0-9c79-17a5be09ec80 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1168.582034] env[63202]: DEBUG oslo_concurrency.lockutils [None req-15c786ed-599e-44c0-9c79-17a5be09ec80 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.618s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1168.599394] env[63202]: INFO nova.scheduler.client.report [None req-15c786ed-599e-44c0-9c79-17a5be09ec80 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Deleted allocations for instance 9646810e-06d2-4de0-8f0c-e8719541da53 [ 1169.106887] env[63202]: DEBUG oslo_concurrency.lockutils [None req-15c786ed-599e-44c0-9c79-17a5be09ec80 tempest-ServersNegativeTestJSON-326092276 tempest-ServersNegativeTestJSON-326092276-project-member] Lock "9646810e-06d2-4de0-8f0c-e8719541da53" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.017s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1174.548697] env[63202]: DEBUG oslo_concurrency.lockutils [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquiring lock "ed60fb5e-b08f-4846-942d-c4edde39d1ae" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1174.549096] env[63202]: DEBUG oslo_concurrency.lockutils [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lock "ed60fb5e-b08f-4846-942d-c4edde39d1ae" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1174.549181] env[63202]: INFO nova.compute.manager [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Shelving [ 1175.055868] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1175.056146] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-336f5d08-9219-4452-8cad-cd7c16681e2f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.064468] env[63202]: DEBUG oslo_vmware.api [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1175.064468] env[63202]: value = "task-1385997" [ 1175.064468] env[63202]: _type = "Task" [ 1175.064468] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.072377] env[63202]: DEBUG oslo_vmware.api [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385997, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.575244] env[63202]: DEBUG oslo_vmware.api [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385997, 'name': PowerOffVM_Task, 'duration_secs': 0.18603} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.575602] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1175.576303] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2c49b01-3f46-4f89-a1bf-71d0e163e0bc {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.594456] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc18cbdd-59e4-4d6e-a28c-91267501cb4b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.105060] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Creating Snapshot of the VM instance {{(pid=63202) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1176.105060] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-0e0d2249-3c3e-4d23-92ec-810abfa866f7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.112224] env[63202]: DEBUG oslo_vmware.api [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1176.112224] env[63202]: value = "task-1385998" [ 1176.112224] env[63202]: _type = "Task" [ 1176.112224] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.119822] env[63202]: DEBUG oslo_vmware.api [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385998, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.622451] env[63202]: DEBUG oslo_vmware.api [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385998, 'name': CreateSnapshot_Task, 'duration_secs': 0.449964} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.622846] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Created Snapshot of the VM instance {{(pid=63202) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1176.623412] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72abe5bb-2661-479c-8c5c-17e5edbd3025 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.140353] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Creating linked-clone VM from snapshot {{(pid=63202) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1177.140665] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-500d1761-f072-4378-a3de-fe5c24f5a6e5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.149480] env[63202]: DEBUG oslo_vmware.api [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1177.149480] env[63202]: value = "task-1385999" [ 1177.149480] env[63202]: _type = "Task" [ 1177.149480] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.157154] env[63202]: DEBUG oslo_vmware.api [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385999, 'name': CloneVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.660041] env[63202]: DEBUG oslo_vmware.api [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385999, 'name': CloneVM_Task} progress is 94%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.159383] env[63202]: DEBUG oslo_vmware.api [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1385999, 'name': CloneVM_Task, 'duration_secs': 1.000939} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.159635] env[63202]: INFO nova.virt.vmwareapi.vmops [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Created linked-clone VM from snapshot [ 1178.160386] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7587a1e-2f61-4e07-a906-53d2dbc364f4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.167225] env[63202]: DEBUG nova.virt.vmwareapi.images [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Uploading image b70d0154-5e2b-4590-8aea-4173ce5347f7 {{(pid=63202) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1178.285547] env[63202]: DEBUG oslo_vmware.rw_handles [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1178.285547] env[63202]: value = "vm-294237" [ 1178.285547] env[63202]: _type = "VirtualMachine" [ 1178.285547] env[63202]: }. {{(pid=63202) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1178.285848] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-412cc2c4-1496-40a6-b31c-f301f1f7134b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.293452] env[63202]: DEBUG oslo_vmware.rw_handles [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lease: (returnval){ [ 1178.293452] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52d4ed70-0149-ace8-2e70-138a80e25e48" [ 1178.293452] env[63202]: _type = "HttpNfcLease" [ 1178.293452] env[63202]: } obtained for exporting VM: (result){ [ 1178.293452] env[63202]: value = "vm-294237" [ 1178.293452] env[63202]: _type = "VirtualMachine" [ 1178.293452] env[63202]: }. {{(pid=63202) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1178.293693] env[63202]: DEBUG oslo_vmware.api [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the lease: (returnval){ [ 1178.293693] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52d4ed70-0149-ace8-2e70-138a80e25e48" [ 1178.293693] env[63202]: _type = "HttpNfcLease" [ 1178.293693] env[63202]: } to be ready. {{(pid=63202) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1178.299765] env[63202]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1178.299765] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52d4ed70-0149-ace8-2e70-138a80e25e48" [ 1178.299765] env[63202]: _type = "HttpNfcLease" [ 1178.299765] env[63202]: } is initializing. {{(pid=63202) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1178.801305] env[63202]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1178.801305] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52d4ed70-0149-ace8-2e70-138a80e25e48" [ 1178.801305] env[63202]: _type = "HttpNfcLease" [ 1178.801305] env[63202]: } is ready. {{(pid=63202) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1178.801750] env[63202]: DEBUG oslo_vmware.rw_handles [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1178.801750] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52d4ed70-0149-ace8-2e70-138a80e25e48" [ 1178.801750] env[63202]: _type = "HttpNfcLease" [ 1178.801750] env[63202]: }. {{(pid=63202) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1178.802235] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-212b1cc0-198b-4bea-868a-e7dd5bf3e03e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.809072] env[63202]: DEBUG oslo_vmware.rw_handles [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52d71559-de82-7e70-97d2-5f2a92d58b85/disk-0.vmdk from lease info. {{(pid=63202) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1178.809243] env[63202]: DEBUG oslo_vmware.rw_handles [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52d71559-de82-7e70-97d2-5f2a92d58b85/disk-0.vmdk for reading. {{(pid=63202) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1178.892030] env[63202]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-9904e4b7-4557-408d-bd28-b155729e6932 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.526480] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1185.526697] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1185.526863] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1185.527013] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1185.587193] env[63202]: DEBUG oslo_vmware.rw_handles [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52d71559-de82-7e70-97d2-5f2a92d58b85/disk-0.vmdk. {{(pid=63202) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1185.588160] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49d4194e-bd37-458f-b542-baed11fb38f8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.594205] env[63202]: DEBUG oslo_vmware.rw_handles [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52d71559-de82-7e70-97d2-5f2a92d58b85/disk-0.vmdk is in state: ready. {{(pid=63202) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1185.594372] env[63202]: ERROR oslo_vmware.rw_handles [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52d71559-de82-7e70-97d2-5f2a92d58b85/disk-0.vmdk due to incomplete transfer. [ 1185.594595] env[63202]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-cd02f642-9fa1-4219-8ba1-ffbfbe6fd270 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.601791] env[63202]: DEBUG oslo_vmware.rw_handles [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52d71559-de82-7e70-97d2-5f2a92d58b85/disk-0.vmdk. {{(pid=63202) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1185.601980] env[63202]: DEBUG nova.virt.vmwareapi.images [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Uploaded image b70d0154-5e2b-4590-8aea-4173ce5347f7 to the Glance image server {{(pid=63202) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1185.604419] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Destroying the VM {{(pid=63202) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1185.604643] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-00e78e68-f7fe-4936-9669-f65cc9b7a942 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.610802] env[63202]: DEBUG oslo_vmware.api [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1185.610802] env[63202]: value = "task-1386001" [ 1185.610802] env[63202]: _type = "Task" [ 1185.610802] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.618256] env[63202]: DEBUG oslo_vmware.api [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1386001, 'name': Destroy_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.120937] env[63202]: DEBUG oslo_vmware.api [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1386001, 'name': Destroy_Task} progress is 100%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.253723] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager.update_available_resource {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1186.621730] env[63202]: DEBUG oslo_vmware.api [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1386001, 'name': Destroy_Task} progress is 100%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.757128] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1186.757356] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1186.757525] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1186.757680] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63202) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1186.758634] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7eb4844e-4418-49b3-bab4-8b74e46c3ddd {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.766543] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-874ed28c-13ab-48bb-b7cf-6821a5234cdf {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.780150] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c9e1beb-e0f9-4b43-9344-d6c00182a533 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.786374] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63a26f90-c61e-4e61-a416-0cf22fff4c64 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.814123] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181261MB free_disk=176GB free_vcpus=48 pci_devices=None {{(pid=63202) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1186.814267] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1186.814456] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1187.121555] env[63202]: DEBUG oslo_vmware.api [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1386001, 'name': Destroy_Task, 'duration_secs': 1.289626} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.121806] env[63202]: INFO nova.virt.vmwareapi.vm_util [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Destroyed the VM [ 1187.122054] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Deleting Snapshot of the VM instance {{(pid=63202) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1187.122302] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-58b4fd80-2ee8-4b00-9cd8-c90f44f7301a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.127762] env[63202]: DEBUG oslo_vmware.api [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1187.127762] env[63202]: value = "task-1386002" [ 1187.127762] env[63202]: _type = "Task" [ 1187.127762] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.135840] env[63202]: DEBUG oslo_vmware.api [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1386002, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.636782] env[63202]: DEBUG oslo_vmware.api [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1386002, 'name': RemoveSnapshot_Task, 'duration_secs': 0.497877} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.638055] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Deleted Snapshot of the VM instance {{(pid=63202) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1187.638055] env[63202]: DEBUG nova.compute.manager [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1187.638644] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d8c15bf-d6ce-417b-9374-62e4969d4935 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.839299] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance ed60fb5e-b08f-4846-942d-c4edde39d1ae actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1187.839453] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1187.839626] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=63202) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1187.839766] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=63202) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1187.875229] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4de2dd6-7f62-466e-8ab0-035be4ab5e94 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.882532] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0811e25e-b082-4bbc-9237-c3f79653ad9d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.910975] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6b0903e-7e7f-4a56-a62c-90f8c6b5698a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.917339] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd0a6bf2-a334-4f04-9aa7-33ee918c03cd {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.929515] env[63202]: DEBUG nova.compute.provider_tree [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1188.150599] env[63202]: INFO nova.compute.manager [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Shelve offloading [ 1188.152171] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1188.152412] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0640d2ec-aa3d-4729-b167-be45cb4e03a9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.159777] env[63202]: DEBUG oslo_vmware.api [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1188.159777] env[63202]: value = "task-1386003" [ 1188.159777] env[63202]: _type = "Task" [ 1188.159777] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.166946] env[63202]: DEBUG oslo_vmware.api [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1386003, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.432988] env[63202]: DEBUG nova.scheduler.client.report [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1188.670088] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] VM already powered off {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1188.670430] env[63202]: DEBUG nova.compute.manager [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1188.671052] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f45822c-d14d-44e0-be89-2050ec0fc141 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.676438] env[63202]: DEBUG oslo_concurrency.lockutils [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquiring lock "refresh_cache-ed60fb5e-b08f-4846-942d-c4edde39d1ae" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1188.676599] env[63202]: DEBUG oslo_concurrency.lockutils [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquired lock "refresh_cache-ed60fb5e-b08f-4846-942d-c4edde39d1ae" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1188.676767] env[63202]: DEBUG nova.network.neutron [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1188.937825] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63202) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1188.938099] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.124s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1189.465706] env[63202]: DEBUG nova.network.neutron [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Updating instance_info_cache with network_info: [{"id": "0193eadb-322b-4278-b259-d56b23555561", "address": "fa:16:3e:1e:87:93", "network": {"id": "06ab5813-9ad9-4021-9bdb-f2f02af8d73f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1714653503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.182", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b2de35030a9484094e964ffc30a822d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3093647a-bab7-4562-ada0-428725e8c0fc", "external-id": "nsx-vlan-transportzone-660", "segmentation_id": 660, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0193eadb-32", "ovs_interfaceid": "0193eadb-322b-4278-b259-d56b23555561", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1189.939053] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1189.939053] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1189.939465] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1189.939465] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63202) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1189.968148] env[63202]: DEBUG oslo_concurrency.lockutils [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Releasing lock "refresh_cache-ed60fb5e-b08f-4846-942d-c4edde39d1ae" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1190.220597] env[63202]: DEBUG nova.compute.manager [req-25be0199-1cd2-4688-91d5-c60f3609e8d1 req-8e563055-6590-42e2-9078-1301ce5b415f service nova] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Received event network-vif-unplugged-0193eadb-322b-4278-b259-d56b23555561 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1190.220826] env[63202]: DEBUG oslo_concurrency.lockutils [req-25be0199-1cd2-4688-91d5-c60f3609e8d1 req-8e563055-6590-42e2-9078-1301ce5b415f service nova] Acquiring lock "ed60fb5e-b08f-4846-942d-c4edde39d1ae-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1190.221147] env[63202]: DEBUG oslo_concurrency.lockutils [req-25be0199-1cd2-4688-91d5-c60f3609e8d1 req-8e563055-6590-42e2-9078-1301ce5b415f service nova] Lock "ed60fb5e-b08f-4846-942d-c4edde39d1ae-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1190.221310] env[63202]: DEBUG oslo_concurrency.lockutils [req-25be0199-1cd2-4688-91d5-c60f3609e8d1 req-8e563055-6590-42e2-9078-1301ce5b415f service nova] Lock "ed60fb5e-b08f-4846-942d-c4edde39d1ae-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1190.221486] env[63202]: DEBUG nova.compute.manager [req-25be0199-1cd2-4688-91d5-c60f3609e8d1 req-8e563055-6590-42e2-9078-1301ce5b415f service nova] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] No waiting events found dispatching network-vif-unplugged-0193eadb-322b-4278-b259-d56b23555561 {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1190.221655] env[63202]: WARNING nova.compute.manager [req-25be0199-1cd2-4688-91d5-c60f3609e8d1 req-8e563055-6590-42e2-9078-1301ce5b415f service nova] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Received unexpected event network-vif-unplugged-0193eadb-322b-4278-b259-d56b23555561 for instance with vm_state shelved and task_state shelving_offloading. [ 1190.279249] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1190.280166] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1afbffda-9a3e-46e4-877f-1a74c800b5ad {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.288566] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1190.288797] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f32b3c0f-99e7-4ab1-9a83-79b8b41c2f32 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.352499] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1190.352738] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Deleting contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1190.352892] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Deleting the datastore file [datastore1] ed60fb5e-b08f-4846-942d-c4edde39d1ae {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1190.353573] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d2168b85-5d3b-4031-8012-80a94ea9072c {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.359656] env[63202]: DEBUG oslo_vmware.api [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1190.359656] env[63202]: value = "task-1386005" [ 1190.359656] env[63202]: _type = "Task" [ 1190.359656] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.367279] env[63202]: DEBUG oslo_vmware.api [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1386005, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.869796] env[63202]: DEBUG oslo_vmware.api [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1386005, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.132534} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.870102] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1190.870289] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Deleted contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1190.870460] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1190.893532] env[63202]: INFO nova.scheduler.client.report [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Deleted allocations for instance ed60fb5e-b08f-4846-942d-c4edde39d1ae [ 1191.398554] env[63202]: DEBUG oslo_concurrency.lockutils [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1191.398935] env[63202]: DEBUG oslo_concurrency.lockutils [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1191.399055] env[63202]: DEBUG nova.objects.instance [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lazy-loading 'resources' on Instance uuid ed60fb5e-b08f-4846-942d-c4edde39d1ae {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1191.901544] env[63202]: DEBUG nova.objects.instance [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lazy-loading 'numa_topology' on Instance uuid ed60fb5e-b08f-4846-942d-c4edde39d1ae {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1192.245245] env[63202]: DEBUG nova.compute.manager [req-2c5ea187-6cd1-455b-b6aa-59616ded98bc req-31146c0e-97b0-46be-964f-fbd878403954 service nova] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Received event network-changed-0193eadb-322b-4278-b259-d56b23555561 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1192.245447] env[63202]: DEBUG nova.compute.manager [req-2c5ea187-6cd1-455b-b6aa-59616ded98bc req-31146c0e-97b0-46be-964f-fbd878403954 service nova] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Refreshing instance network info cache due to event network-changed-0193eadb-322b-4278-b259-d56b23555561. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1192.245657] env[63202]: DEBUG oslo_concurrency.lockutils [req-2c5ea187-6cd1-455b-b6aa-59616ded98bc req-31146c0e-97b0-46be-964f-fbd878403954 service nova] Acquiring lock "refresh_cache-ed60fb5e-b08f-4846-942d-c4edde39d1ae" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1192.245800] env[63202]: DEBUG oslo_concurrency.lockutils [req-2c5ea187-6cd1-455b-b6aa-59616ded98bc req-31146c0e-97b0-46be-964f-fbd878403954 service nova] Acquired lock "refresh_cache-ed60fb5e-b08f-4846-942d-c4edde39d1ae" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1192.245961] env[63202]: DEBUG nova.network.neutron [req-2c5ea187-6cd1-455b-b6aa-59616ded98bc req-31146c0e-97b0-46be-964f-fbd878403954 service nova] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Refreshing network info cache for port 0193eadb-322b-4278-b259-d56b23555561 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1192.404148] env[63202]: DEBUG nova.objects.base [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=63202) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1192.430592] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e101459f-9d80-405e-a827-22c36c07ca4e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.440397] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-519bb1fa-3e14-4523-9d0d-c28e18ef57e8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.485191] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-306de4dd-ded5-44b4-b09a-5fa24835aedc {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.492077] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f47ab7a5-5331-41d0-9b70-75c9deb8a95b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.504400] env[63202]: DEBUG nova.compute.provider_tree [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1193.007861] env[63202]: DEBUG nova.scheduler.client.report [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1193.153169] env[63202]: DEBUG nova.network.neutron [req-2c5ea187-6cd1-455b-b6aa-59616ded98bc req-31146c0e-97b0-46be-964f-fbd878403954 service nova] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Updated VIF entry in instance network info cache for port 0193eadb-322b-4278-b259-d56b23555561. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1193.153550] env[63202]: DEBUG nova.network.neutron [req-2c5ea187-6cd1-455b-b6aa-59616ded98bc req-31146c0e-97b0-46be-964f-fbd878403954 service nova] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Updating instance_info_cache with network_info: [{"id": "0193eadb-322b-4278-b259-d56b23555561", "address": "fa:16:3e:1e:87:93", "network": {"id": "06ab5813-9ad9-4021-9bdb-f2f02af8d73f", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1714653503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.182", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b2de35030a9484094e964ffc30a822d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap0193eadb-32", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1193.254726] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1193.254884] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Starting heal instance info cache {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1193.255000] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Rebuilding the list of instances to heal {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1193.512274] env[63202]: DEBUG oslo_concurrency.lockutils [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.113s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1193.656218] env[63202]: DEBUG oslo_concurrency.lockutils [req-2c5ea187-6cd1-455b-b6aa-59616ded98bc req-31146c0e-97b0-46be-964f-fbd878403954 service nova] Releasing lock "refresh_cache-ed60fb5e-b08f-4846-942d-c4edde39d1ae" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1193.758456] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Acquiring lock "refresh_cache-ed60fb5e-b08f-4846-942d-c4edde39d1ae" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1193.758585] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Acquired lock "refresh_cache-ed60fb5e-b08f-4846-942d-c4edde39d1ae" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1193.758735] env[63202]: DEBUG nova.network.neutron [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Forcefully refreshing network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1193.758880] env[63202]: DEBUG nova.objects.instance [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lazy-loading 'info_cache' on Instance uuid ed60fb5e-b08f-4846-942d-c4edde39d1ae {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1193.868017] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3e0f90c8-ea66-4883-b89e-7410feab0cf2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquiring lock "0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1193.868017] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3e0f90c8-ea66-4883-b89e-7410feab0cf2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1193.982468] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquiring lock "ed60fb5e-b08f-4846-942d-c4edde39d1ae" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1194.019610] env[63202]: DEBUG oslo_concurrency.lockutils [None req-515f27d7-2413-4b5c-be28-4234378e7ef4 tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lock "ed60fb5e-b08f-4846-942d-c4edde39d1ae" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 19.471s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1194.021514] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lock "ed60fb5e-b08f-4846-942d-c4edde39d1ae" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.038s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1194.021514] env[63202]: INFO nova.compute.manager [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Unshelving [ 1194.372068] env[63202]: DEBUG nova.compute.utils [None req-3e0f90c8-ea66-4883-b89e-7410feab0cf2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Using /dev/sd instead of None {{(pid=63202) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1194.874343] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3e0f90c8-ea66-4883-b89e-7410feab0cf2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1195.043844] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1195.044241] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1195.044477] env[63202]: DEBUG nova.objects.instance [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lazy-loading 'pci_requests' on Instance uuid ed60fb5e-b08f-4846-942d-c4edde39d1ae {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1195.548708] env[63202]: DEBUG nova.objects.instance [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lazy-loading 'numa_topology' on Instance uuid ed60fb5e-b08f-4846-942d-c4edde39d1ae {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1195.578968] env[63202]: DEBUG nova.network.neutron [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Updating instance_info_cache with network_info: [{"id": "0193eadb-322b-4278-b259-d56b23555561", "address": "fa:16:3e:1e:87:93", "network": {"id": "06ab5813-9ad9-4021-9bdb-f2f02af8d73f", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1714653503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.182", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b2de35030a9484094e964ffc30a822d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap0193eadb-32", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1195.942830] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3e0f90c8-ea66-4883-b89e-7410feab0cf2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquiring lock "0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1195.943284] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3e0f90c8-ea66-4883-b89e-7410feab0cf2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1195.943396] env[63202]: INFO nova.compute.manager [None req-3e0f90c8-ea66-4883-b89e-7410feab0cf2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Attaching volume 3f5f5592-b8c2-4ce8-a84e-008f6b385dcd to /dev/sdb [ 1195.974767] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56cfb0cd-3f36-46a3-ae62-0890d10dc5c2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.982031] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c8172e4-ada1-41b9-b8b0-590d5751212e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.996305] env[63202]: DEBUG nova.virt.block_device [None req-3e0f90c8-ea66-4883-b89e-7410feab0cf2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Updating existing volume attachment record: 2095a31f-aaff-4172-b97a-8e84e4735cd3 {{(pid=63202) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1196.050960] env[63202]: INFO nova.compute.claims [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1196.081165] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Releasing lock "refresh_cache-ed60fb5e-b08f-4846-942d-c4edde39d1ae" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1196.081360] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Updated the network info_cache for instance {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1197.096203] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b79110c-9d26-4142-bbe7-6df4efc068e5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.103482] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14636562-58cf-4f97-beb1-419c2f6a9f93 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.132287] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35b9f92c-ce48-4fef-aa36-43d6e57153ba {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.139105] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c1656e1-0e2c-42f1-b903-c80b76e63714 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.151407] env[63202]: DEBUG nova.compute.provider_tree [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1197.654445] env[63202]: DEBUG nova.scheduler.client.report [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1198.159041] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.115s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1198.186524] env[63202]: INFO nova.network.neutron [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Updating port 0193eadb-322b-4278-b259-d56b23555561 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1199.569444] env[63202]: DEBUG nova.compute.manager [req-176d10bd-3017-4746-81ac-47a14b1968cb req-74e9bf34-0799-466a-b421-b49f197416b9 service nova] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Received event network-vif-plugged-0193eadb-322b-4278-b259-d56b23555561 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1199.569698] env[63202]: DEBUG oslo_concurrency.lockutils [req-176d10bd-3017-4746-81ac-47a14b1968cb req-74e9bf34-0799-466a-b421-b49f197416b9 service nova] Acquiring lock "ed60fb5e-b08f-4846-942d-c4edde39d1ae-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1199.569863] env[63202]: DEBUG oslo_concurrency.lockutils [req-176d10bd-3017-4746-81ac-47a14b1968cb req-74e9bf34-0799-466a-b421-b49f197416b9 service nova] Lock "ed60fb5e-b08f-4846-942d-c4edde39d1ae-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1199.571895] env[63202]: DEBUG oslo_concurrency.lockutils [req-176d10bd-3017-4746-81ac-47a14b1968cb req-74e9bf34-0799-466a-b421-b49f197416b9 service nova] Lock "ed60fb5e-b08f-4846-942d-c4edde39d1ae-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1199.571895] env[63202]: DEBUG nova.compute.manager [req-176d10bd-3017-4746-81ac-47a14b1968cb req-74e9bf34-0799-466a-b421-b49f197416b9 service nova] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] No waiting events found dispatching network-vif-plugged-0193eadb-322b-4278-b259-d56b23555561 {{(pid=63202) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1199.571895] env[63202]: WARNING nova.compute.manager [req-176d10bd-3017-4746-81ac-47a14b1968cb req-74e9bf34-0799-466a-b421-b49f197416b9 service nova] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Received unexpected event network-vif-plugged-0193eadb-322b-4278-b259-d56b23555561 for instance with vm_state shelved_offloaded and task_state spawning. [ 1199.664613] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquiring lock "refresh_cache-ed60fb5e-b08f-4846-942d-c4edde39d1ae" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1199.664613] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquired lock "refresh_cache-ed60fb5e-b08f-4846-942d-c4edde39d1ae" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1199.664613] env[63202]: DEBUG nova.network.neutron [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Building network info cache for instance {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1200.365493] env[63202]: DEBUG nova.network.neutron [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Updating instance_info_cache with network_info: [{"id": "0193eadb-322b-4278-b259-d56b23555561", "address": "fa:16:3e:1e:87:93", "network": {"id": "06ab5813-9ad9-4021-9bdb-f2f02af8d73f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1714653503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.182", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b2de35030a9484094e964ffc30a822d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3093647a-bab7-4562-ada0-428725e8c0fc", "external-id": "nsx-vlan-transportzone-660", "segmentation_id": 660, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0193eadb-32", "ovs_interfaceid": "0193eadb-322b-4278-b259-d56b23555561", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1200.542054] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e0f90c8-ea66-4883-b89e-7410feab0cf2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Volume attach. Driver type: vmdk {{(pid=63202) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1200.542369] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e0f90c8-ea66-4883-b89e-7410feab0cf2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294238', 'volume_id': '3f5f5592-b8c2-4ce8-a84e-008f6b385dcd', 'name': 'volume-3f5f5592-b8c2-4ce8-a84e-008f6b385dcd', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8', 'attached_at': '', 'detached_at': '', 'volume_id': '3f5f5592-b8c2-4ce8-a84e-008f6b385dcd', 'serial': '3f5f5592-b8c2-4ce8-a84e-008f6b385dcd'} {{(pid=63202) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1200.543730] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15663185-e1d9-4776-9164-ac8aedf2ac13 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.559731] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24c6ac59-750e-45c9-94ac-9b88fbc092f3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.583840] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e0f90c8-ea66-4883-b89e-7410feab0cf2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] volume-3f5f5592-b8c2-4ce8-a84e-008f6b385dcd/volume-3f5f5592-b8c2-4ce8-a84e-008f6b385dcd.vmdk or device None with type thin {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1200.584185] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5871c345-c1bc-4fa2-ac3e-0c9f1c689efe {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.602161] env[63202]: DEBUG oslo_vmware.api [None req-3e0f90c8-ea66-4883-b89e-7410feab0cf2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 1200.602161] env[63202]: value = "task-1386008" [ 1200.602161] env[63202]: _type = "Task" [ 1200.602161] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.609434] env[63202]: DEBUG oslo_vmware.api [None req-3e0f90c8-ea66-4883-b89e-7410feab0cf2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1386008, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.868604] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Releasing lock "refresh_cache-ed60fb5e-b08f-4846-942d-c4edde39d1ae" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1200.897624] env[63202]: DEBUG nova.virt.hardware [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T10:20:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='dc12864bd5e6fc86aea66a8bcfaebcf1',container_format='bare',created_at=2024-10-31T10:32:38Z,direct_url=,disk_format='vmdk',id=b70d0154-5e2b-4590-8aea-4173ce5347f7,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-431775784-shelved',owner='2b2de35030a9484094e964ffc30a822d',properties=ImageMetaProps,protected=,size=31670272,status='active',tags=,updated_at=2024-10-31T10:32:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1200.897894] env[63202]: DEBUG nova.virt.hardware [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Flavor limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1200.898109] env[63202]: DEBUG nova.virt.hardware [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Image limits 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1200.898324] env[63202]: DEBUG nova.virt.hardware [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Flavor pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1200.898489] env[63202]: DEBUG nova.virt.hardware [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Image pref 0:0:0 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1200.898643] env[63202]: DEBUG nova.virt.hardware [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63202) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1200.898849] env[63202]: DEBUG nova.virt.hardware [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1200.899041] env[63202]: DEBUG nova.virt.hardware [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1200.899235] env[63202]: DEBUG nova.virt.hardware [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Got 1 possible topologies {{(pid=63202) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1200.899415] env[63202]: DEBUG nova.virt.hardware [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1200.899603] env[63202]: DEBUG nova.virt.hardware [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63202) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1200.900546] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-470774a5-7ab2-429c-9a6f-9ae306319b0a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.908413] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcaad2c3-d511-4ea5-b279-0dd26a6e5bcf {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.922266] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1e:87:93', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3093647a-bab7-4562-ada0-428725e8c0fc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0193eadb-322b-4278-b259-d56b23555561', 'vif_model': 'vmxnet3'}] {{(pid=63202) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1200.930775] env[63202]: DEBUG oslo.service.loopingcall [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1200.931046] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Creating VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1200.931276] env[63202]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-33d7b7e8-6399-446f-8300-9761978d79bf {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.949748] env[63202]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1200.949748] env[63202]: value = "task-1386009" [ 1200.949748] env[63202]: _type = "Task" [ 1200.949748] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.962295] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1386009, 'name': CreateVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.112150] env[63202]: DEBUG oslo_vmware.api [None req-3e0f90c8-ea66-4883-b89e-7410feab0cf2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1386008, 'name': ReconfigVM_Task, 'duration_secs': 0.366476} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1201.112453] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e0f90c8-ea66-4883-b89e-7410feab0cf2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Reconfigured VM instance instance-0000006b to attach disk [datastore1] volume-3f5f5592-b8c2-4ce8-a84e-008f6b385dcd/volume-3f5f5592-b8c2-4ce8-a84e-008f6b385dcd.vmdk or device None with type thin {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1201.117298] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-07a81b9a-0e89-477b-b622-f9897b7f1515 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.131880] env[63202]: DEBUG oslo_vmware.api [None req-3e0f90c8-ea66-4883-b89e-7410feab0cf2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 1201.131880] env[63202]: value = "task-1386010" [ 1201.131880] env[63202]: _type = "Task" [ 1201.131880] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1201.140909] env[63202]: DEBUG oslo_vmware.api [None req-3e0f90c8-ea66-4883-b89e-7410feab0cf2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1386010, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.458883] env[63202]: DEBUG oslo_vmware.api [-] Task: {'id': task-1386009, 'name': CreateVM_Task, 'duration_secs': 0.305822} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1201.459098] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Created VM on the ESX host {{(pid=63202) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1201.459704] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b70d0154-5e2b-4590-8aea-4173ce5347f7" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1201.459873] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b70d0154-5e2b-4590-8aea-4173ce5347f7" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1201.460285] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b70d0154-5e2b-4590-8aea-4173ce5347f7" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1201.460563] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a7b300f0-52f4-463f-b8af-565fcea31dbe {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.464918] env[63202]: DEBUG oslo_vmware.api [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1201.464918] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52775fad-b107-26e2-b051-238a52b3ff33" [ 1201.464918] env[63202]: _type = "Task" [ 1201.464918] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1201.471951] env[63202]: DEBUG oslo_vmware.api [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52775fad-b107-26e2-b051-238a52b3ff33, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.593715] env[63202]: DEBUG nova.compute.manager [req-5d9b270d-dc64-445a-b4d1-cb0409a9cbf7 req-d665cda2-a76a-459c-b6e4-fac6f886f724 service nova] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Received event network-changed-0193eadb-322b-4278-b259-d56b23555561 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1201.594200] env[63202]: DEBUG nova.compute.manager [req-5d9b270d-dc64-445a-b4d1-cb0409a9cbf7 req-d665cda2-a76a-459c-b6e4-fac6f886f724 service nova] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Refreshing instance network info cache due to event network-changed-0193eadb-322b-4278-b259-d56b23555561. {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1201.594578] env[63202]: DEBUG oslo_concurrency.lockutils [req-5d9b270d-dc64-445a-b4d1-cb0409a9cbf7 req-d665cda2-a76a-459c-b6e4-fac6f886f724 service nova] Acquiring lock "refresh_cache-ed60fb5e-b08f-4846-942d-c4edde39d1ae" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1201.594808] env[63202]: DEBUG oslo_concurrency.lockutils [req-5d9b270d-dc64-445a-b4d1-cb0409a9cbf7 req-d665cda2-a76a-459c-b6e4-fac6f886f724 service nova] Acquired lock "refresh_cache-ed60fb5e-b08f-4846-942d-c4edde39d1ae" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1201.595145] env[63202]: DEBUG nova.network.neutron [req-5d9b270d-dc64-445a-b4d1-cb0409a9cbf7 req-d665cda2-a76a-459c-b6e4-fac6f886f724 service nova] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Refreshing network info cache for port 0193eadb-322b-4278-b259-d56b23555561 {{(pid=63202) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1201.640583] env[63202]: DEBUG oslo_vmware.api [None req-3e0f90c8-ea66-4883-b89e-7410feab0cf2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1386010, 'name': ReconfigVM_Task, 'duration_secs': 0.156454} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1201.640874] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e0f90c8-ea66-4883-b89e-7410feab0cf2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294238', 'volume_id': '3f5f5592-b8c2-4ce8-a84e-008f6b385dcd', 'name': 'volume-3f5f5592-b8c2-4ce8-a84e-008f6b385dcd', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8', 'attached_at': '', 'detached_at': '', 'volume_id': '3f5f5592-b8c2-4ce8-a84e-008f6b385dcd', 'serial': '3f5f5592-b8c2-4ce8-a84e-008f6b385dcd'} {{(pid=63202) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1201.975262] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b70d0154-5e2b-4590-8aea-4173ce5347f7" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1201.975533] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Processing image b70d0154-5e2b-4590-8aea-4173ce5347f7 {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1201.975771] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b70d0154-5e2b-4590-8aea-4173ce5347f7/b70d0154-5e2b-4590-8aea-4173ce5347f7.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1201.975921] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b70d0154-5e2b-4590-8aea-4173ce5347f7/b70d0154-5e2b-4590-8aea-4173ce5347f7.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1201.976117] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1201.976358] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c6f6a590-8160-4a69-bd58-2766e4e17b9a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.992523] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1201.992700] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63202) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1201.993452] env[63202]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e4282183-0ab6-4e53-8586-3dae34135591 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.998941] env[63202]: DEBUG oslo_vmware.api [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1201.998941] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]52c09f11-0d66-5889-5aed-69a3620e8579" [ 1201.998941] env[63202]: _type = "Task" [ 1201.998941] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1202.006351] env[63202]: DEBUG oslo_vmware.api [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': session[52076eab-3258-8204-7106-1fff41a217da]52c09f11-0d66-5889-5aed-69a3620e8579, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.300399] env[63202]: DEBUG nova.network.neutron [req-5d9b270d-dc64-445a-b4d1-cb0409a9cbf7 req-d665cda2-a76a-459c-b6e4-fac6f886f724 service nova] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Updated VIF entry in instance network info cache for port 0193eadb-322b-4278-b259-d56b23555561. {{(pid=63202) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1202.300748] env[63202]: DEBUG nova.network.neutron [req-5d9b270d-dc64-445a-b4d1-cb0409a9cbf7 req-d665cda2-a76a-459c-b6e4-fac6f886f724 service nova] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Updating instance_info_cache with network_info: [{"id": "0193eadb-322b-4278-b259-d56b23555561", "address": "fa:16:3e:1e:87:93", "network": {"id": "06ab5813-9ad9-4021-9bdb-f2f02af8d73f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1714653503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.182", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b2de35030a9484094e964ffc30a822d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3093647a-bab7-4562-ada0-428725e8c0fc", "external-id": "nsx-vlan-transportzone-660", "segmentation_id": 660, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0193eadb-32", "ovs_interfaceid": "0193eadb-322b-4278-b259-d56b23555561", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1202.509249] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Preparing fetch location {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1202.509507] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Fetch image to [datastore1] OSTACK_IMG_e278e830-33a1-4dd2-a9e8-09749e318e47/OSTACK_IMG_e278e830-33a1-4dd2-a9e8-09749e318e47.vmdk {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1202.509703] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Downloading stream optimized image b70d0154-5e2b-4590-8aea-4173ce5347f7 to [datastore1] OSTACK_IMG_e278e830-33a1-4dd2-a9e8-09749e318e47/OSTACK_IMG_e278e830-33a1-4dd2-a9e8-09749e318e47.vmdk on the data store datastore1 as vApp {{(pid=63202) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1202.509880] env[63202]: DEBUG nova.virt.vmwareapi.images [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Downloading image file data b70d0154-5e2b-4590-8aea-4173ce5347f7 to the ESX as VM named 'OSTACK_IMG_e278e830-33a1-4dd2-a9e8-09749e318e47' {{(pid=63202) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1202.573189] env[63202]: DEBUG oslo_vmware.rw_handles [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1202.573189] env[63202]: value = "resgroup-9" [ 1202.573189] env[63202]: _type = "ResourcePool" [ 1202.573189] env[63202]: }. {{(pid=63202) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1202.573499] env[63202]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-a5f5e652-ca6a-40db-a84d-9e282e038f9e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.593427] env[63202]: DEBUG oslo_vmware.rw_handles [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lease: (returnval){ [ 1202.593427] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]522d453f-720e-a233-fdd4-f8ad6604b596" [ 1202.593427] env[63202]: _type = "HttpNfcLease" [ 1202.593427] env[63202]: } obtained for vApp import into resource pool (val){ [ 1202.593427] env[63202]: value = "resgroup-9" [ 1202.593427] env[63202]: _type = "ResourcePool" [ 1202.593427] env[63202]: }. {{(pid=63202) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1202.593702] env[63202]: DEBUG oslo_vmware.api [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the lease: (returnval){ [ 1202.593702] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]522d453f-720e-a233-fdd4-f8ad6604b596" [ 1202.593702] env[63202]: _type = "HttpNfcLease" [ 1202.593702] env[63202]: } to be ready. {{(pid=63202) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1202.600697] env[63202]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1202.600697] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]522d453f-720e-a233-fdd4-f8ad6604b596" [ 1202.600697] env[63202]: _type = "HttpNfcLease" [ 1202.600697] env[63202]: } is initializing. {{(pid=63202) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1202.674198] env[63202]: DEBUG nova.objects.instance [None req-3e0f90c8-ea66-4883-b89e-7410feab0cf2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lazy-loading 'flavor' on Instance uuid 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1202.803468] env[63202]: DEBUG oslo_concurrency.lockutils [req-5d9b270d-dc64-445a-b4d1-cb0409a9cbf7 req-d665cda2-a76a-459c-b6e4-fac6f886f724 service nova] Releasing lock "refresh_cache-ed60fb5e-b08f-4846-942d-c4edde39d1ae" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1203.102392] env[63202]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1203.102392] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]522d453f-720e-a233-fdd4-f8ad6604b596" [ 1203.102392] env[63202]: _type = "HttpNfcLease" [ 1203.102392] env[63202]: } is ready. {{(pid=63202) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1203.102695] env[63202]: DEBUG oslo_vmware.rw_handles [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1203.102695] env[63202]: value = "session[52076eab-3258-8204-7106-1fff41a217da]522d453f-720e-a233-fdd4-f8ad6604b596" [ 1203.102695] env[63202]: _type = "HttpNfcLease" [ 1203.102695] env[63202]: }. {{(pid=63202) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1203.103441] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c396615-7397-4bc7-acde-3a50f5e81ca3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.110131] env[63202]: DEBUG oslo_vmware.rw_handles [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b9faa1-f996-2ae7-990a-6eff9e75461b/disk-0.vmdk from lease info. {{(pid=63202) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1203.110315] env[63202]: DEBUG oslo_vmware.rw_handles [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Creating HTTP connection to write to file with size = 31670272 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b9faa1-f996-2ae7-990a-6eff9e75461b/disk-0.vmdk. {{(pid=63202) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1203.176151] env[63202]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b4c0731b-708c-4f7a-b3d3-46473522ef4b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.180481] env[63202]: DEBUG oslo_concurrency.lockutils [None req-3e0f90c8-ea66-4883-b89e-7410feab0cf2 tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.237s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1203.357386] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5277c53a-90e6-475e-8dad-fed9152c640f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquiring lock "0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1203.357711] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5277c53a-90e6-475e-8dad-fed9152c640f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1203.860581] env[63202]: INFO nova.compute.manager [None req-5277c53a-90e6-475e-8dad-fed9152c640f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Detaching volume 3f5f5592-b8c2-4ce8-a84e-008f6b385dcd [ 1203.894512] env[63202]: INFO nova.virt.block_device [None req-5277c53a-90e6-475e-8dad-fed9152c640f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Attempting to driver detach volume 3f5f5592-b8c2-4ce8-a84e-008f6b385dcd from mountpoint /dev/sdb [ 1203.894824] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-5277c53a-90e6-475e-8dad-fed9152c640f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Volume detach. Driver type: vmdk {{(pid=63202) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1203.895073] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-5277c53a-90e6-475e-8dad-fed9152c640f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294238', 'volume_id': '3f5f5592-b8c2-4ce8-a84e-008f6b385dcd', 'name': 'volume-3f5f5592-b8c2-4ce8-a84e-008f6b385dcd', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8', 'attached_at': '', 'detached_at': '', 'volume_id': '3f5f5592-b8c2-4ce8-a84e-008f6b385dcd', 'serial': '3f5f5592-b8c2-4ce8-a84e-008f6b385dcd'} {{(pid=63202) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1203.896068] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78bef417-cabf-4cc5-afe3-c56e91853972 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.924898] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-903bd977-cc78-47a6-8da8-d5ec6d27d4c3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.934786] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-313750f4-9434-43aa-913f-abeb9f614d24 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.959101] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eec4958a-c907-465b-a229-49bd5d0e6892 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.979394] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-5277c53a-90e6-475e-8dad-fed9152c640f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] The volume has not been displaced from its original location: [datastore1] volume-3f5f5592-b8c2-4ce8-a84e-008f6b385dcd/volume-3f5f5592-b8c2-4ce8-a84e-008f6b385dcd.vmdk. No consolidation needed. {{(pid=63202) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1203.988073] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-5277c53a-90e6-475e-8dad-fed9152c640f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Reconfiguring VM instance instance-0000006b to detach disk 2001 {{(pid=63202) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1203.988547] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-62160b59-0540-4398-b176-11a9ae50444b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.008897] env[63202]: DEBUG oslo_vmware.api [None req-5277c53a-90e6-475e-8dad-fed9152c640f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 1204.008897] env[63202]: value = "task-1386012" [ 1204.008897] env[63202]: _type = "Task" [ 1204.008897] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1204.016992] env[63202]: DEBUG oslo_vmware.api [None req-5277c53a-90e6-475e-8dad-fed9152c640f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1386012, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.215998] env[63202]: DEBUG oslo_vmware.rw_handles [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Completed reading data from the image iterator. {{(pid=63202) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1204.216240] env[63202]: DEBUG oslo_vmware.rw_handles [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b9faa1-f996-2ae7-990a-6eff9e75461b/disk-0.vmdk. {{(pid=63202) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1204.217272] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf6499a0-e306-42c6-b615-748e66662a2d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.223781] env[63202]: DEBUG oslo_vmware.rw_handles [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b9faa1-f996-2ae7-990a-6eff9e75461b/disk-0.vmdk is in state: ready. {{(pid=63202) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1204.223954] env[63202]: DEBUG oslo_vmware.rw_handles [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b9faa1-f996-2ae7-990a-6eff9e75461b/disk-0.vmdk. {{(pid=63202) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1204.224216] env[63202]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-292e100c-6953-44ad-9995-eaa605fbed85 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.439880] env[63202]: DEBUG oslo_vmware.rw_handles [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b9faa1-f996-2ae7-990a-6eff9e75461b/disk-0.vmdk. {{(pid=63202) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1204.440127] env[63202]: INFO nova.virt.vmwareapi.images [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Downloaded image file data b70d0154-5e2b-4590-8aea-4173ce5347f7 [ 1204.441031] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2a894c0-0dd3-42b3-bfab-ca05319c2b3f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.457433] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3f5274c6-4313-4da8-ba60-fc9670d3f9a8 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.488028] env[63202]: INFO nova.virt.vmwareapi.images [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] The imported VM was unregistered [ 1204.490341] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Caching image {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1204.490589] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Creating directory with path [datastore1] devstack-image-cache_base/b70d0154-5e2b-4590-8aea-4173ce5347f7 {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1204.490845] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8156c92f-3a33-4f2e-99cc-d990d0c4d324 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.500016] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Created directory with path [datastore1] devstack-image-cache_base/b70d0154-5e2b-4590-8aea-4173ce5347f7 {{(pid=63202) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1204.500216] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_e278e830-33a1-4dd2-a9e8-09749e318e47/OSTACK_IMG_e278e830-33a1-4dd2-a9e8-09749e318e47.vmdk to [datastore1] devstack-image-cache_base/b70d0154-5e2b-4590-8aea-4173ce5347f7/b70d0154-5e2b-4590-8aea-4173ce5347f7.vmdk. {{(pid=63202) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1204.500445] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-7ffc6b50-c39f-4dc4-9362-e5648fc66338 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.505876] env[63202]: DEBUG oslo_vmware.api [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1204.505876] env[63202]: value = "task-1386014" [ 1204.505876] env[63202]: _type = "Task" [ 1204.505876] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1204.514735] env[63202]: DEBUG oslo_vmware.api [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1386014, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.519015] env[63202]: DEBUG oslo_vmware.api [None req-5277c53a-90e6-475e-8dad-fed9152c640f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1386012, 'name': ReconfigVM_Task, 'duration_secs': 0.257198} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1204.519276] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-5277c53a-90e6-475e-8dad-fed9152c640f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Reconfigured VM instance instance-0000006b to detach disk 2001 {{(pid=63202) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1204.523828] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-67ba3693-5a63-4657-93a2-9f7d2fce8af4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.537304] env[63202]: DEBUG oslo_vmware.api [None req-5277c53a-90e6-475e-8dad-fed9152c640f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 1204.537304] env[63202]: value = "task-1386015" [ 1204.537304] env[63202]: _type = "Task" [ 1204.537304] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1204.544355] env[63202]: DEBUG oslo_vmware.api [None req-5277c53a-90e6-475e-8dad-fed9152c640f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1386015, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.018336] env[63202]: DEBUG oslo_vmware.api [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1386014, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.049634] env[63202]: DEBUG oslo_vmware.api [None req-5277c53a-90e6-475e-8dad-fed9152c640f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1386015, 'name': ReconfigVM_Task, 'duration_secs': 0.132374} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1205.049978] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-5277c53a-90e6-475e-8dad-fed9152c640f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294238', 'volume_id': '3f5f5592-b8c2-4ce8-a84e-008f6b385dcd', 'name': 'volume-3f5f5592-b8c2-4ce8-a84e-008f6b385dcd', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8', 'attached_at': '', 'detached_at': '', 'volume_id': '3f5f5592-b8c2-4ce8-a84e-008f6b385dcd', 'serial': '3f5f5592-b8c2-4ce8-a84e-008f6b385dcd'} {{(pid=63202) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1205.520231] env[63202]: DEBUG oslo_vmware.api [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1386014, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.597180] env[63202]: DEBUG nova.objects.instance [None req-5277c53a-90e6-475e-8dad-fed9152c640f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lazy-loading 'flavor' on Instance uuid 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1206.020550] env[63202]: DEBUG oslo_vmware.api [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1386014, 'name': MoveVirtualDisk_Task} progress is 69%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.521405] env[63202]: DEBUG oslo_vmware.api [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1386014, 'name': MoveVirtualDisk_Task} progress is 94%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.605252] env[63202]: DEBUG oslo_concurrency.lockutils [None req-5277c53a-90e6-475e-8dad-fed9152c640f tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.247s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1207.018291] env[63202]: DEBUG oslo_vmware.api [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1386014, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.179162} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1207.018559] env[63202]: INFO nova.virt.vmwareapi.ds_util [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_e278e830-33a1-4dd2-a9e8-09749e318e47/OSTACK_IMG_e278e830-33a1-4dd2-a9e8-09749e318e47.vmdk to [datastore1] devstack-image-cache_base/b70d0154-5e2b-4590-8aea-4173ce5347f7/b70d0154-5e2b-4590-8aea-4173ce5347f7.vmdk. [ 1207.018748] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Cleaning up location [datastore1] OSTACK_IMG_e278e830-33a1-4dd2-a9e8-09749e318e47 {{(pid=63202) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1207.018914] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_e278e830-33a1-4dd2-a9e8-09749e318e47 {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1207.019181] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8c033170-48aa-4120-a2d5-509992e59c90 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.025224] env[63202]: DEBUG oslo_vmware.api [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1207.025224] env[63202]: value = "task-1386016" [ 1207.025224] env[63202]: _type = "Task" [ 1207.025224] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1207.032282] env[63202]: DEBUG oslo_vmware.api [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1386016, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.535164] env[63202]: DEBUG oslo_vmware.api [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1386016, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.031689} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1207.535413] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1207.535595] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b70d0154-5e2b-4590-8aea-4173ce5347f7/b70d0154-5e2b-4590-8aea-4173ce5347f7.vmdk" {{(pid=63202) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1207.535858] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b70d0154-5e2b-4590-8aea-4173ce5347f7/b70d0154-5e2b-4590-8aea-4173ce5347f7.vmdk to [datastore1] ed60fb5e-b08f-4846-942d-c4edde39d1ae/ed60fb5e-b08f-4846-942d-c4edde39d1ae.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1207.536120] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-facea8b6-7420-475e-964d-a53f28273293 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.542577] env[63202]: DEBUG oslo_vmware.api [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1207.542577] env[63202]: value = "task-1386017" [ 1207.542577] env[63202]: _type = "Task" [ 1207.542577] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1207.549877] env[63202]: DEBUG oslo_vmware.api [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1386017, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.611596] env[63202]: DEBUG oslo_concurrency.lockutils [None req-74df8135-dca4-476d-9f6d-bf6d3e70984e tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquiring lock "0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1207.611790] env[63202]: DEBUG oslo_concurrency.lockutils [None req-74df8135-dca4-476d-9f6d-bf6d3e70984e tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1207.611996] env[63202]: DEBUG oslo_concurrency.lockutils [None req-74df8135-dca4-476d-9f6d-bf6d3e70984e tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquiring lock "0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1207.612198] env[63202]: DEBUG oslo_concurrency.lockutils [None req-74df8135-dca4-476d-9f6d-bf6d3e70984e tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1207.612372] env[63202]: DEBUG oslo_concurrency.lockutils [None req-74df8135-dca4-476d-9f6d-bf6d3e70984e tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1207.614553] env[63202]: INFO nova.compute.manager [None req-74df8135-dca4-476d-9f6d-bf6d3e70984e tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Terminating instance [ 1207.616458] env[63202]: DEBUG nova.compute.manager [None req-74df8135-dca4-476d-9f6d-bf6d3e70984e tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1207.616650] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-74df8135-dca4-476d-9f6d-bf6d3e70984e tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1207.617441] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64040d72-6351-40c0-8e6d-32a861df7dc1 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.624156] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-74df8135-dca4-476d-9f6d-bf6d3e70984e tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1207.624386] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1d92b8ca-ef99-4f3d-951e-a0e98b9b3e1f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.630529] env[63202]: DEBUG oslo_vmware.api [None req-74df8135-dca4-476d-9f6d-bf6d3e70984e tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 1207.630529] env[63202]: value = "task-1386018" [ 1207.630529] env[63202]: _type = "Task" [ 1207.630529] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1207.637450] env[63202]: DEBUG oslo_vmware.api [None req-74df8135-dca4-476d-9f6d-bf6d3e70984e tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1386018, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.057101] env[63202]: DEBUG oslo_vmware.api [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1386017, 'name': CopyVirtualDisk_Task} progress is 18%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.144861] env[63202]: DEBUG oslo_vmware.api [None req-74df8135-dca4-476d-9f6d-bf6d3e70984e tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1386018, 'name': PowerOffVM_Task, 'duration_secs': 0.200325} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1208.145194] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-74df8135-dca4-476d-9f6d-bf6d3e70984e tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1208.145384] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-74df8135-dca4-476d-9f6d-bf6d3e70984e tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1208.145660] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-57f64183-efef-4f8a-8a20-0c8f31ebd405 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.226943] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-74df8135-dca4-476d-9f6d-bf6d3e70984e tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1208.227280] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-74df8135-dca4-476d-9f6d-bf6d3e70984e tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Deleting contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1208.227492] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-74df8135-dca4-476d-9f6d-bf6d3e70984e tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Deleting the datastore file [datastore1] 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8 {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1208.227815] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c3812a39-8f17-4330-87ff-157d691f6f0e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.239552] env[63202]: DEBUG oslo_vmware.api [None req-74df8135-dca4-476d-9f6d-bf6d3e70984e tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for the task: (returnval){ [ 1208.239552] env[63202]: value = "task-1386020" [ 1208.239552] env[63202]: _type = "Task" [ 1208.239552] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1208.252885] env[63202]: DEBUG oslo_vmware.api [None req-74df8135-dca4-476d-9f6d-bf6d3e70984e tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1386020, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.556338] env[63202]: DEBUG oslo_vmware.api [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1386017, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.753412] env[63202]: DEBUG oslo_vmware.api [None req-74df8135-dca4-476d-9f6d-bf6d3e70984e tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1386020, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.059598] env[63202]: DEBUG oslo_vmware.api [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1386017, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.253592] env[63202]: DEBUG oslo_vmware.api [None req-74df8135-dca4-476d-9f6d-bf6d3e70984e tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1386020, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.561704] env[63202]: DEBUG oslo_vmware.api [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1386017, 'name': CopyVirtualDisk_Task} progress is 85%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.754325] env[63202]: DEBUG oslo_vmware.api [None req-74df8135-dca4-476d-9f6d-bf6d3e70984e tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1386020, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.055948] env[63202]: DEBUG oslo_vmware.api [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1386017, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.312361} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1210.056231] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b70d0154-5e2b-4590-8aea-4173ce5347f7/b70d0154-5e2b-4590-8aea-4173ce5347f7.vmdk to [datastore1] ed60fb5e-b08f-4846-942d-c4edde39d1ae/ed60fb5e-b08f-4846-942d-c4edde39d1ae.vmdk {{(pid=63202) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1210.057058] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eb37c54-32f3-4832-a2b0-3156f61c7208 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.078635] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] ed60fb5e-b08f-4846-942d-c4edde39d1ae/ed60fb5e-b08f-4846-942d-c4edde39d1ae.vmdk or device None with type streamOptimized {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1210.079045] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-441e6dd0-1a67-4bf7-9b92-904e68657d74 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.097146] env[63202]: DEBUG oslo_vmware.api [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1210.097146] env[63202]: value = "task-1386021" [ 1210.097146] env[63202]: _type = "Task" [ 1210.097146] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1210.105724] env[63202]: DEBUG oslo_vmware.api [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1386021, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.252096] env[63202]: DEBUG oslo_vmware.api [None req-74df8135-dca4-476d-9f6d-bf6d3e70984e tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Task: {'id': task-1386020, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.610478} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1210.252395] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-74df8135-dca4-476d-9f6d-bf6d3e70984e tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1210.252589] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-74df8135-dca4-476d-9f6d-bf6d3e70984e tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Deleted contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1210.252768] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-74df8135-dca4-476d-9f6d-bf6d3e70984e tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1210.252998] env[63202]: INFO nova.compute.manager [None req-74df8135-dca4-476d-9f6d-bf6d3e70984e tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Took 2.64 seconds to destroy the instance on the hypervisor. [ 1210.253288] env[63202]: DEBUG oslo.service.loopingcall [None req-74df8135-dca4-476d-9f6d-bf6d3e70984e tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1210.253486] env[63202]: DEBUG nova.compute.manager [-] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1210.253610] env[63202]: DEBUG nova.network.neutron [-] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1210.607834] env[63202]: DEBUG oslo_vmware.api [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1386021, 'name': ReconfigVM_Task, 'duration_secs': 0.297643} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1210.608116] env[63202]: DEBUG nova.virt.vmwareapi.volumeops [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Reconfigured VM instance instance-0000006a to attach disk [datastore1] ed60fb5e-b08f-4846-942d-c4edde39d1ae/ed60fb5e-b08f-4846-942d-c4edde39d1ae.vmdk or device None with type streamOptimized {{(pid=63202) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1210.608780] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ccc4938f-31d9-4fe0-a212-4ed5f33b36f6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.614765] env[63202]: DEBUG oslo_vmware.api [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1210.614765] env[63202]: value = "task-1386022" [ 1210.614765] env[63202]: _type = "Task" [ 1210.614765] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1210.622430] env[63202]: DEBUG oslo_vmware.api [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1386022, 'name': Rename_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.705273] env[63202]: DEBUG nova.compute.manager [req-99078a5f-f5d9-4518-8395-bb3287effbe7 req-d461d69c-077a-440f-b319-8bf657321dbe service nova] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Received event network-vif-deleted-5a9c7dc7-9386-4f78-89ab-720b71c72161 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1210.705443] env[63202]: INFO nova.compute.manager [req-99078a5f-f5d9-4518-8395-bb3287effbe7 req-d461d69c-077a-440f-b319-8bf657321dbe service nova] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Neutron deleted interface 5a9c7dc7-9386-4f78-89ab-720b71c72161; detaching it from the instance and deleting it from the info cache [ 1210.705631] env[63202]: DEBUG nova.network.neutron [req-99078a5f-f5d9-4518-8395-bb3287effbe7 req-d461d69c-077a-440f-b319-8bf657321dbe service nova] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1211.125110] env[63202]: DEBUG oslo_vmware.api [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1386022, 'name': Rename_Task, 'duration_secs': 0.133874} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1211.125110] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Powering on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1211.125591] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b998a8fa-5130-4feb-9d5f-574c2f1a43e2 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.130770] env[63202]: DEBUG oslo_vmware.api [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1211.130770] env[63202]: value = "task-1386023" [ 1211.130770] env[63202]: _type = "Task" [ 1211.130770] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1211.137598] env[63202]: DEBUG oslo_vmware.api [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1386023, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.184126] env[63202]: DEBUG nova.network.neutron [-] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1211.207633] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-100f1182-3d72-41a6-8a8b-3d1bb92ff1a6 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.216965] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97091cbe-a747-4fb7-8a81-477d3e56e1ea {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.241608] env[63202]: DEBUG nova.compute.manager [req-99078a5f-f5d9-4518-8395-bb3287effbe7 req-d461d69c-077a-440f-b319-8bf657321dbe service nova] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Detach interface failed, port_id=5a9c7dc7-9386-4f78-89ab-720b71c72161, reason: Instance 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8 could not be found. {{(pid=63202) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1211.641710] env[63202]: DEBUG oslo_vmware.api [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1386023, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.686434] env[63202]: INFO nova.compute.manager [-] [instance: 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8] Took 1.43 seconds to deallocate network for instance. [ 1212.142500] env[63202]: DEBUG oslo_vmware.api [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1386023, 'name': PowerOnVM_Task, 'duration_secs': 0.550698} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1212.142843] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Powered on the VM {{(pid=63202) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1212.192700] env[63202]: DEBUG oslo_concurrency.lockutils [None req-74df8135-dca4-476d-9f6d-bf6d3e70984e tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1212.192992] env[63202]: DEBUG oslo_concurrency.lockutils [None req-74df8135-dca4-476d-9f6d-bf6d3e70984e tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1212.193270] env[63202]: DEBUG nova.objects.instance [None req-74df8135-dca4-476d-9f6d-bf6d3e70984e tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lazy-loading 'resources' on Instance uuid 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8 {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1212.239109] env[63202]: DEBUG nova.compute.manager [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Checking state {{(pid=63202) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1212.240378] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33feebbd-4415-4acc-bcb1-594768d9937e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.735389] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a07c3f1c-6127-47c1-a62a-1b31b8c9a810 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.742637] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10c509d0-9d58-4188-b1cc-95b08edcab1e {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.775794] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22c6b23b-d460-4ae4-9b41-c5401b61f3b9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.778276] env[63202]: DEBUG oslo_concurrency.lockutils [None req-c973abf4-99da-43ab-9d41-18dc9f98dd5b tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lock "ed60fb5e-b08f-4846-942d-c4edde39d1ae" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 18.758s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1212.783868] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aab1512a-5ecc-4666-b0ff-b59145060f8a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.798040] env[63202]: DEBUG nova.compute.provider_tree [None req-74df8135-dca4-476d-9f6d-bf6d3e70984e tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1213.300574] env[63202]: DEBUG nova.scheduler.client.report [None req-74df8135-dca4-476d-9f6d-bf6d3e70984e tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1213.806297] env[63202]: DEBUG oslo_concurrency.lockutils [None req-74df8135-dca4-476d-9f6d-bf6d3e70984e tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.613s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1213.825621] env[63202]: INFO nova.scheduler.client.report [None req-74df8135-dca4-476d-9f6d-bf6d3e70984e tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Deleted allocations for instance 0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8 [ 1214.335605] env[63202]: DEBUG oslo_concurrency.lockutils [None req-74df8135-dca4-476d-9f6d-bf6d3e70984e tempest-AttachVolumeNegativeTest-1243065399 tempest-AttachVolumeNegativeTest-1243065399-project-member] Lock "0c1b85f5-0d2a-41bf-acec-d1f6bfc6cca8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.724s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1244.254155] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1244.254605] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1245.255149] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1246.249498] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1247.254427] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager.update_available_resource {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1247.757064] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1247.757339] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1247.757506] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1247.757661] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63202) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1247.758591] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47c6393d-aa7d-4e3b-b5ef-a36abf6ba118 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.766803] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-816d7bba-c4fc-432b-8392-988e2d36b2a4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.779823] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55daec53-ca9f-499b-be5a-f2228e4b5268 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.785795] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e958da68-3d7a-4fb3-8504-c98bc7f4c30b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.814332] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181195MB free_disk=176GB free_vcpus=48 pci_devices=None {{(pid=63202) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1247.814470] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1247.814681] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1247.862910] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7a535c9d-3122-459f-8a97-bff3b5a9a12d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquiring lock "ed60fb5e-b08f-4846-942d-c4edde39d1ae" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1247.863171] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7a535c9d-3122-459f-8a97-bff3b5a9a12d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lock "ed60fb5e-b08f-4846-942d-c4edde39d1ae" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1247.863373] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7a535c9d-3122-459f-8a97-bff3b5a9a12d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquiring lock "ed60fb5e-b08f-4846-942d-c4edde39d1ae-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1247.863555] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7a535c9d-3122-459f-8a97-bff3b5a9a12d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lock "ed60fb5e-b08f-4846-942d-c4edde39d1ae-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1247.863775] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7a535c9d-3122-459f-8a97-bff3b5a9a12d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lock "ed60fb5e-b08f-4846-942d-c4edde39d1ae-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1247.866693] env[63202]: INFO nova.compute.manager [None req-7a535c9d-3122-459f-8a97-bff3b5a9a12d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Terminating instance [ 1247.868426] env[63202]: DEBUG nova.compute.manager [None req-7a535c9d-3122-459f-8a97-bff3b5a9a12d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Start destroying the instance on the hypervisor. {{(pid=63202) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1247.868619] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-7a535c9d-3122-459f-8a97-bff3b5a9a12d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Destroying instance {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1247.869413] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97a44064-4633-45ea-9d71-0b8249421de4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.877231] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a535c9d-3122-459f-8a97-bff3b5a9a12d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Powering off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1247.877444] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-64f52f07-d74a-42fb-afa4-e834ff30c747 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.884722] env[63202]: DEBUG oslo_vmware.api [None req-7a535c9d-3122-459f-8a97-bff3b5a9a12d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1247.884722] env[63202]: value = "task-1386027" [ 1247.884722] env[63202]: _type = "Task" [ 1247.884722] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1247.891708] env[63202]: DEBUG oslo_vmware.api [None req-7a535c9d-3122-459f-8a97-bff3b5a9a12d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1386027, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1248.395696] env[63202]: DEBUG oslo_vmware.api [None req-7a535c9d-3122-459f-8a97-bff3b5a9a12d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1386027, 'name': PowerOffVM_Task, 'duration_secs': 0.192029} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1248.396080] env[63202]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a535c9d-3122-459f-8a97-bff3b5a9a12d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Powered off the VM {{(pid=63202) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1248.396164] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-7a535c9d-3122-459f-8a97-bff3b5a9a12d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Unregistering the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1248.396422] env[63202]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-894d46b1-d82b-470f-b9ad-c3d1b3816416 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.457854] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-7a535c9d-3122-459f-8a97-bff3b5a9a12d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Unregistered the VM {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1248.458146] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-7a535c9d-3122-459f-8a97-bff3b5a9a12d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Deleting contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1248.458345] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a535c9d-3122-459f-8a97-bff3b5a9a12d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Deleting the datastore file [datastore1] ed60fb5e-b08f-4846-942d-c4edde39d1ae {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1248.458593] env[63202]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7a1b8355-4d7d-49d6-9ea3-01880eb1c5f9 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.466080] env[63202]: DEBUG oslo_vmware.api [None req-7a535c9d-3122-459f-8a97-bff3b5a9a12d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for the task: (returnval){ [ 1248.466080] env[63202]: value = "task-1386029" [ 1248.466080] env[63202]: _type = "Task" [ 1248.466080] env[63202]: } to complete. {{(pid=63202) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1248.473485] env[63202]: DEBUG oslo_vmware.api [None req-7a535c9d-3122-459f-8a97-bff3b5a9a12d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1386029, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1248.838939] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Instance ed60fb5e-b08f-4846-942d-c4edde39d1ae actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63202) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1248.838939] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=63202) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1248.838939] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=63202) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1248.862765] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52f1d0c1-d3b8-4564-a4fb-9c82264db4c5 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.870188] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b82bfcc-4ef7-4622-a92d-d91694b3fb2b {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.898481] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-213dbf8d-db92-4e12-937a-ea05b5efb3f7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.905119] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7440d1c-cc73-4754-8e0a-37c97e07235f {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.918312] env[63202]: DEBUG nova.compute.provider_tree [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1248.975156] env[63202]: DEBUG oslo_vmware.api [None req-7a535c9d-3122-459f-8a97-bff3b5a9a12d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Task: {'id': task-1386029, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.121905} completed successfully. {{(pid=63202) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1248.975393] env[63202]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a535c9d-3122-459f-8a97-bff3b5a9a12d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Deleted the datastore file {{(pid=63202) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1248.975572] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-7a535c9d-3122-459f-8a97-bff3b5a9a12d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Deleted contents of the VM from datastore datastore1 {{(pid=63202) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1248.975746] env[63202]: DEBUG nova.virt.vmwareapi.vmops [None req-7a535c9d-3122-459f-8a97-bff3b5a9a12d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Instance destroyed {{(pid=63202) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1248.975916] env[63202]: INFO nova.compute.manager [None req-7a535c9d-3122-459f-8a97-bff3b5a9a12d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1248.976166] env[63202]: DEBUG oslo.service.loopingcall [None req-7a535c9d-3122-459f-8a97-bff3b5a9a12d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63202) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1248.976358] env[63202]: DEBUG nova.compute.manager [-] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Deallocating network for instance {{(pid=63202) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1248.976448] env[63202]: DEBUG nova.network.neutron [-] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] deallocate_for_instance() {{(pid=63202) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1249.420919] env[63202]: DEBUG nova.scheduler.client.report [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1249.426089] env[63202]: DEBUG nova.compute.manager [req-283a9e94-e957-46be-b0e6-827d8309979e req-56b2211d-c8cd-4df1-8a8f-e15bd38319b8 service nova] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Received event network-vif-deleted-0193eadb-322b-4278-b259-d56b23555561 {{(pid=63202) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1249.426292] env[63202]: INFO nova.compute.manager [req-283a9e94-e957-46be-b0e6-827d8309979e req-56b2211d-c8cd-4df1-8a8f-e15bd38319b8 service nova] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Neutron deleted interface 0193eadb-322b-4278-b259-d56b23555561; detaching it from the instance and deleting it from the info cache [ 1249.426461] env[63202]: DEBUG nova.network.neutron [req-283a9e94-e957-46be-b0e6-827d8309979e req-56b2211d-c8cd-4df1-8a8f-e15bd38319b8 service nova] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1249.899347] env[63202]: DEBUG nova.network.neutron [-] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Updating instance_info_cache with network_info: [] {{(pid=63202) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1249.929015] env[63202]: DEBUG nova.compute.resource_tracker [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63202) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1249.929230] env[63202]: DEBUG oslo_concurrency.lockutils [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.115s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1249.929852] env[63202]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-85212822-0120-4b80-ac1a-28592bcd03b3 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.940994] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-625ac266-8183-4fc1-b306-8e1005d1bf94 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.962887] env[63202]: DEBUG nova.compute.manager [req-283a9e94-e957-46be-b0e6-827d8309979e req-56b2211d-c8cd-4df1-8a8f-e15bd38319b8 service nova] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Detach interface failed, port_id=0193eadb-322b-4278-b259-d56b23555561, reason: Instance ed60fb5e-b08f-4846-942d-c4edde39d1ae could not be found. {{(pid=63202) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1250.402431] env[63202]: INFO nova.compute.manager [-] [instance: ed60fb5e-b08f-4846-942d-c4edde39d1ae] Took 1.43 seconds to deallocate network for instance. [ 1250.908879] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7a535c9d-3122-459f-8a97-bff3b5a9a12d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1250.909227] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7a535c9d-3122-459f-8a97-bff3b5a9a12d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1250.909407] env[63202]: DEBUG nova.objects.instance [None req-7a535c9d-3122-459f-8a97-bff3b5a9a12d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lazy-loading 'resources' on Instance uuid ed60fb5e-b08f-4846-942d-c4edde39d1ae {{(pid=63202) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1250.928712] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1250.928885] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1250.929082] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1250.929204] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63202) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1251.249313] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1251.443071] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb02502a-13dc-4b27-8fbc-efb6b233d7f7 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.450702] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b7916fc-69c7-414a-995b-33cdf82feec4 {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.480441] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53dd668c-2c5b-403e-b50a-83011fa1359d {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.487081] env[63202]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72bb12a3-15b0-45bf-b1b5-8759f43e6c0a {{(pid=63202) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.499718] env[63202]: DEBUG nova.compute.provider_tree [None req-7a535c9d-3122-459f-8a97-bff3b5a9a12d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Inventory has not changed in ProviderTree for provider: 79b33d17-3e75-494c-a550-67b275de2079 {{(pid=63202) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1252.003258] env[63202]: DEBUG nova.scheduler.client.report [None req-7a535c9d-3122-459f-8a97-bff3b5a9a12d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Inventory has not changed for provider 79b33d17-3e75-494c-a550-67b275de2079 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63202) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1252.508502] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7a535c9d-3122-459f-8a97-bff3b5a9a12d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.599s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1252.528670] env[63202]: INFO nova.scheduler.client.report [None req-7a535c9d-3122-459f-8a97-bff3b5a9a12d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Deleted allocations for instance ed60fb5e-b08f-4846-942d-c4edde39d1ae [ 1253.036826] env[63202]: DEBUG oslo_concurrency.lockutils [None req-7a535c9d-3122-459f-8a97-bff3b5a9a12d tempest-AttachVolumeShelveTestJSON-977664705 tempest-AttachVolumeShelveTestJSON-977664705-project-member] Lock "ed60fb5e-b08f-4846-942d-c4edde39d1ae" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.173s {{(pid=63202) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1254.254177] env[63202]: DEBUG oslo_service.periodic_task [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63202) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1254.254402] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Starting heal instance info cache {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1254.757404] env[63202]: DEBUG nova.compute.manager [None req-f76687af-0467-48f3-9524-cc997e4d720f None None] Didn't find any instances for network info cache update. {{(pid=63202) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}}